ZhangZhihui's Blog  

2025年2月11日

摘要: zzh@ZZHPC:~$ which pip /home/zzh/venvs/zpy313/bin/pip zzh@ZZHPC:~$ pip install chromadb Collecting chromadb Downloading chromadb-0.6.3-py3-none-any.wh 阅读全文
posted @ 2025-02-11 12:06 ZhangZhihuiAAA 阅读(2) 评论(0) 推荐(0) 编辑
 
摘要: Analyze and Optimize sqlite_stat1 is an internal table. It is not in the output of .tables. Suggested Pragmas Faster inserts Method 1 (a bit risky): M 阅读全文
posted @ 2025-02-11 10:39 ZhangZhihuiAAA 阅读(1) 评论(0) 推荐(0) 编辑

2025年2月10日

摘要: Flexible Typing Type is at cell level. Sqlite will convert the type to the declared type if it can without losing data. 5 data types: Type Affinity Co 阅读全文
posted @ 2025-02-10 16:01 ZhangZhihuiAAA 阅读(2) 评论(0) 推荐(0) 编辑
 
摘要: zzh@ZZHPC:~$ sudo apt install sqlite3 zzh@ZZHPC:~$ sqlite3 SQLite version 3.37.2 2022-01-06 13:25:41 Enter ".help" for usage hints. Connected to a tra 阅读全文
posted @ 2025-02-10 13:29 ZhangZhihuiAAA 阅读(2) 评论(0) 推荐(0) 编辑

2025年2月9日

摘要: In Databricks Community Edition, you cannot use Workflows because it is a premium feature that requires an upgraded subscription. This chapter and the 阅读全文
posted @ 2025-02-09 21:02 ZhangZhihuiAAA 阅读(2) 评论(0) 推荐(0) 编辑
 
摘要: from delta import configure_spark_with_delta_pip from pyspark.sql import SparkSession from pyspark.sql.functions import when, rand import timeit build 阅读全文
posted @ 2025-02-09 16:09 ZhangZhihuiAAA 阅读(2) 评论(0) 推荐(0) 编辑

2025年2月8日

摘要: from pyspark.sql import SparkSession # Create a new SparkSession spark = (SparkSession .builder .appName("monitor-spark-ui") .master("spark://ZZHPC:70 阅读全文
posted @ 2025-02-08 13:15 ZhangZhihuiAAA 阅读(2) 评论(0) 推荐(0) 编辑

2025年2月7日

摘要: from delta import configure_spark_with_delta_pip, DeltaTable from pyspark.sql import SparkSession from pyspark.sql.functions import col, from_json fro 阅读全文
posted @ 2025-02-07 18:21 ZhangZhihuiAAA 阅读(1) 评论(0) 推荐(0) 编辑

2025年2月5日

摘要: nc -lk 9999 from pyspark.sql import SparkSession from pyspark.sql.functions import explode, split spark = (SparkSession.builder .appName("config-strea 阅读全文
posted @ 2025-02-05 16:34 ZhangZhihuiAAA 阅读(2) 评论(0) 推荐(0) 编辑

2025年2月3日

摘要: 1. Download Spark 3.4.1 2. Download Java JDK 17 3. Setup Python virtual environment 3.11.9 .bashrc: sfw=~/Downloads/sfw zpy=~/venvs/zpy311 export JAVA 阅读全文
posted @ 2025-02-03 17:33 ZhangZhihuiAAA 阅读(4) 评论(0) 推荐(0) 编辑