flipboard.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
Welcome to Flipboard on Mastodon. A place for our community of curators and enthusiasts to inform and inspire each other. If you'd like to join please request an invitation via the sign-up page.

Administered by:

Server stats:

1.2K
active users

#clickhouse

3 posts3 participants0 posts today

🥱 #ClickHouse just discovered that being #lazy is lucrative and they're throwing a #party to celebrate—because who doesn’t want to attend a #conference for a product that's too lazy to load its own data? 🐌 Apparently, they’re also offering a "Bring Your Own Cloud" feature, so you can experience the thrill of paying for something you already own! 🎉
clickhouse.com/blog/clickhouse #BYO #Cloud #Data #Celebration #HackerNews #ngated

ClickHouseClickHouse gets lazier (and faster): Introducing lazy materializationClickHouse learned to procrastinate strategically. Discover how lazy materialization skips unnecessary column reads to accelerate queries.

🌖 ClickHouse 的 Rust 一年
➤ 從實驗到實用,ClickHouse 如何擁抱 Rust
clickhouse.com/blog/rust
ClickHouse 團隊決定嘗試將 Rust 整合到其以 C++ 撰寫的程式碼庫中,並非為了完全重寫,而是為了吸引 Rust 工程師,並利用 Rust 在特定領域的優勢。他們從整合 BLAKE3 哈希函數開始,之後又加入了 PRQL 查詢語言和 Delta Lake 支援。儘管過程中遇到一些挑戰,例如效能問題和供應鏈問題,但整體而言,Rust 的整合為 ClickHouse 帶來了新的可能性,並提高了團隊對 Rust 的信心。
+ 很有趣的案例!證明瞭即使是大型 C++ 專案,也可以逐步導入其他語言,並從中獲益。
+ 期待 ClickHouse 能夠持續整合更多 Rust 程式碼,讓更多 Rust 工程師參與到這個專案中。
#開源專案 #資料庫 #Rust #ClickHouse

ClickHouseA Year of Rust in ClickHouseThis story is about how ClickHouse supports Rust components in the C++ code base and the challenges we had to overcome.

oh man there's a TUI for #clickHouse called #chdig and i really wish i knew about this a couple of years ago when i was working with a couple of teams that used clickhouse a lot because i was really curious what it looked like in there. from the chdig README there's a chart of tools and layers of abstraction for clickhouse that make a tool like this (now in #macOS #homebrew `brew install chdig`) especially useful.

Continued thread

Okay, there -- first pass done. github.com/scottlaird/nel-coll is a #NEL (Network Error Log) collector that can listen on HTTP for NEL messages from Chrome, etc and log them to a DB.

It should be at least somewhat DB-agnostic, although I've only tested it with #clickhouse so far.

It also includes otel tracing, which is uglier than I'd like in Go, but seems to be working. I still need to get trace propagation through it to the DB backend working, but I think that's a problem with the backend config, not my code. Probably.

I'm still using #vector for this at home right now; I'll probably cut over tomorrow.

Join Alkin Tezuysal's session "Unified Observability: Leveraging ClickHouse as a Comprehensive Telemetry Database" at #FOSSASIASummit2025 to learn how #ClickHouse simplifies observability, giving engineers deep system insights while minimizing infrastructure complexity.

🔗 Click here youtu.be/sCRXGqOPtOU?si=lWcjSC to watch on the FOSSASIA YouTube channel.

youtu.be- YouTubeEnjoy the videos and music you love, upload original content, and share it all with friends, family, and the world on YouTube.

Unraveling the Mystery of CPU Spikes in ClickHouse Cloud: A Deep Dive into Linux Kernel Debugging

A perplexing CPU spike in ClickHouse Cloud led to an extensive investigation, revealing a hidden bug in the Linux kernel's memory management. This article chronicles the journey through kernel interna...

news.lavx.hu/article/unravelin

In the last couple of weeks (on and off between other things) I was looking into a #ClickHouse based implementation of #Prometheus' rate() functionality on #OpenTelemetry metrics.

There are a couple of vendors aiming to provide this feature, but most of them either do it by counting traces/spans, or using inflexible schemas and/or not handling edge-cases properly.

Stay tuned!

New blog post: how to pull web logs from #Caddy into #Clickhouse using #Vector.

scottstuff.net/posts/2025/02/2

Clickhouse is an open-source (plus paid, as usual) columnar DB. This lets you do ad hoc SQL queries to answer questions as well as create Grafana dashboards to show trends, etc.

scottstuff.net · Getting Caddy logs into Clickhouse via VectorAs mentioned before, I’ve been using the Caddy web server running on a couple machines to serve this site. I’ve been dumping Caddy’s access logs into Grafana’s Loki log system, but I haven’t been very happy with it for web logs. It’s kind of a pain to configure for small uses (a few GB of data on one server), and it’s slow for my use case. I’m sure I could optimize it one way or another, but even without the performance issues I’m still not very happy with it for logs analysis. I’ve had a number of relatively simple queries that I’ve had to fight with both Loki and Grafana to get answers for. In this specific case, I was trying to understand how much traffic my post on the Minisforum MS-A2 was getting and where it was coming from, and it was easier for me to grep through a few GB of gzipped JSON log files than to get Loki to answer my questions. So maybe it’s not the right tool for the job and I should look at other options. I’d been meaning to look at Clickhouse for a while; it’s an open source (plus paid cloud offering) column-store analytical DB. You feed it data and then use SQL to query it. It similar to Google BigQuery, Dremel, etc, and dozens of other similar systems. The big advantage of column-oriented databases is that queries that only hit a few fields can be really fast, because they can ignore all of the other columns completely. So a typical analytic query can just do giant streaming reads from a couple column without any disk seeks, which means your performance mostly just ends up being limited by your disks’ streaming throughput. Not so hot when you want to fetch all of the data from a single record, but great when you want to read millions of rows and calculate aggregate statistics. I managed to get Clickhouse reading Caddy’s logs, but it wasn’t quite as trivial as I’d hoped, and none of the assorted “how to do things like this” docs that I found online really covered this case very well, so I figured I’d write up the process that I used.

🌘 可靠地在PostgreSQL和ClickHouse之間複製數據第一部分 - PeerDB 開源 | BenjaminWootton.com
➤ PostgreSQL和ClickHouse之間的數據複製方法及選擇,以及PeerDB在其中的角色。
benjaminwootton.com/insights/c
本文討論如何從PostgreSQL到ClickHouse可靠複製數據。介紹利用PeerDB開源進行數據複製的方法,隨著ClickHouse在實時數據倉庫中的增長,PostgreSQL和ClickHouse的組合自然地被廣泛應用。
+ 這篇文章清晰地解釋了在PostgreSQL和ClickHouse之間複製數據的流程和工具,對於相關領域的人很有幫助。
+ 精簡明瞭的摘要讓人一目瞭然,對於有興趣瞭解數據複製的讀者來說很寶貴。
#數據複製 #PostgreSQL #ClickHouse

BenjaminWootton.comReliably Replicating Data Between PostgreSQL and ClickHouse Part 1 - PeerDB Open Source | BenjaminWootton.comUsing PeerDB open source to reliably replicate data from PostgreSQL to ClickHouse.

🌘 Clickhouse 大量插入:如何避免過多插入導致問題
➤ Clickhouse 大量插入最佳實踐及工具推薦
runportcullis.co/blog/bulk-dat
2025 年 2 月份進行 Clickhouse 大量插入時,瞭解 MergeTree 如何處理資料合併,將資料批次插入可提升效能。避免太多小型插入,監控系統表如 system.parts 及 system.merges 可調整策略。部署工具如 Jitsu Bulker 和 Clickhouse Bulk 可提升效率。
+ 這篇文章提供了實用的 Clickhouse 大量插入技巧,對處理大數據很有幫助。
+ 詳細介紹了 Clickhouse 的插入機制和最佳實踐,對想要優化資料庫效能的人很有參考價值。
#Clickhouse 大量插入

www.runportcullis.coBulk inserts on Clickhouse: How to avoid overstuffing your instanceAs we hit the midway point of the second month in 2025, a lot of you might be starting to really dig in on new data initiatives and planning key infrastructure changes to your company’s data stack.