Hacker News new | past | comments | ask | show | jobs | submit | lukekim's comments login

Hi HN, we’re Luke and Philip, founders of Spice AI. Today, we’re announcing Spice.ai OSS 1.0-Stable, a portable, single-node data query and LLM-inference engine built in Rust.

We introduced the very first POC of Spice.ai OSS on ShowHN in Sep 2021 as a runtime for building AI-driven applications using time-series data!! [Insert, it’s been 87-years meme].

One of the hard lessons we learned was that before we organizations could use AI effectively, they needed a much higher level of data readiness. Our customers told us they wanted to adopt AI and technologies like Arrow, Iceberg, Delta Lake, and DuckDB but simply didn’t have the time or resources (and were struggling to keep up), so we focused on making it super easy and simple to use them. We rebuilt Spice from the ground up in Rust on Apache DataFusion, and launched on ShowHN in Mar 2024 as a unified SQL query interface to locally materialize, accelerate, and query datasets sourced from any data source.

It’s designed for developers who want to build fast, reliable data-intensive and AI apps without getting stuck managing ETL pipelines or complex infrastructure.

That release was just the data foundation and today, we’re announcing Spice.ai OSS 1.0-Stable that includes federated data query, acceleration, retrieval, and AI inference into a single engine—now ready for production deployments across cloud, BYOC, edge, or on-prem.

Spice supports accelerating federated queries across databases (MySQL, PostgreSQL, etc.), data warehouses (Snowflake, Databricks, etc.), and data lakes (S3, MinIO, etc.). It materializes datasets locally using Arrow, DuckDB, or SQLite for sub-second query times and integrates LLM-inference and memory capabilities and a purpose-built data-grounding toolset that includes vector and hybrid search, Text-to-SQL/NSQL, and evals to ensure accurate outputs.

We’d love for you to check it out on GitHub, try it, and share your feedback: https://github.com/spiceai/spiceai

Thank you!

[1] Spice.ai OSS v0.1 announcement: https://news.ycombinator.com/item?id=28449182

[2] Spice.ai OSS rebuilt in Rust: https://news.ycombinator.com/item?id=39854584


You also have to run it as admin. If you don't, it tries to auto-update every day and when it can't presents a big, obnoxious red banner.

[1] https://x.com/WarpSpeedDan/status/1811447481047097788


Aloke from Warp here. The experience if you don’t have admin access isn’t great—I’ll follow up and make sure this gets fixed.


The Model Context server is similar to what we've built at Spice, but we've focused on databases and data systems. Overall, standards are good. Perhaps we can implement MCP as a data connector and tool.

[1] https://github.com/spiceai/spiceai


We chose Apache 2.0 for the Spice OSS runtime.

TL;DR: Data-plane Apache 2.0, control-plane BSL.

Being such a core component, we want developers to be completely comfortable integrating and deploying the Spice runtime in their applications and services, as well as running Spice in their own infrastructure.

In addition, Spice OSS is built on other great open-source projects like DataFusion and Arrow, both Apache 2.0, and DuckDB (MIT), so being permissively licensed aligns with the fundamental technologies and communities it's built upon.

We expect to release specific enterprise control-plane services, such as our Kubernetes Operator under a license such as BSL.

[1] https://github.com/spiceai/spiceai


Spice AI | SWE & DevRel | FT | ONSITE (Seattle, Seoul), REMOTE (Australia) Spice AI is the creator of the Spice.ai open-source project, a query-engine and ML inferencing runtime built in Rust on DataFusion. Hiring experienced Rust, distributed systems, data systems, and database engineers and DevRel. ShowHN: https://news.ycombinator.com/item?id=39854584 Details: https://spice.ai/careers


Also anecdotal, but we (Spice AI) see more requests for Iceberg, but in practice more deployments of Delta Lake.


My theory is that everyone would _prefer_ to use Iceberg, but isn’t as widely supported _yet_, so they’re stuck with Delta in the interim.


I think you're spot on


Thanks! Feedback and GitHub issues welcome!


Yes, it's on the backlog and we'll prioritize as we see demand as with https://github.com/spiceai/spiceai/issues/999.


Spice supports what DataFusion supports, which is generally yes but there is still work to do to push down more queries to TableProviders. For example, joins within a single source are not yet pushed down to the underlying provider.

You can write a single query across many data sources which is what we show in the demo on the Git repo.


There is an effort within DataFusion to support pushing down joins across tables from the same remote provider that we will likely contribute to as well: https://github.com/datafusion-contrib/datafusion-federation


Thank you!

Yes, in terms of federated queries, there are similarities, but Spice is designed to be much smaller, faster, and lightweight (single-binary, 140MB) so you can run it next to your application as a sidecar, or eventually even in the browser. Spice also gives you more options and flexibility for materialization, so you can choose where and how to store local materialized data.


Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: