Data Reliability at Chick-Fil-A(medium.com) Chick-fil-A has over 3,000 locations across the USA, Puerto Rico, and Canada, with over 8 million orders per day. The amount of data being tracked and processed, including Restaurant data points, customer orders, and other business operations information creates a data rich landscape, but also a multitude of challenges. Data Reliability Engineering (DRE) helps Chick-fil-A approach these challenges and utilize resources to create a reliable system that supports the business and customers on a daily basis.
Federated Data Access for MCP (Model Context Protocol)(mindsdb.com) Today marks a significant milestone in our mission to simplify how AI accesses enterprise data. We're excited to announce that MindsDB now fully supports the Model Context Protocol (MCP) across both our open source and enterprise platforms. This gives our enterprise customers and open source users a unified way for their AI applications and agents to run queries over federated data stored in different databases and clouds as if it were a single database.
Declarative Schemas for simpler database management(supabase.com) Today we’re releasing declarative schemas to simplify managing and maintaining complex database schemas. With declarative schemas, you can define your database structure in a clear, centralized, and version-controlled manner.
Palantir suggests 'common operating system' for UK govt data(theregister.com) In a witness statement to the UK COVID-19 Inquiry [PDF], an ongoing independent public inquiry into the nation's response to the pandemic (in which around 208,000 people died), Louis Mosley, executive veep of Palantir Technologies UK, said the government should invest in a "common operating system" for its data, encompassing departments such as the Department for Work and Pensions and local authorities.
Multiply Went from Datomic to XTDB to Rama(redplanetlabs.com) "With databases, the conversation always started with ‘what are we able to do?’. I rarely find myself asking what Rama is able to support, and rather ‘how?’. The requirements of the application dictate how we utilise the platform, not the other way around. Rama as a tool allows us to think product first, while still delivering highly optimised and scalable features for specific use cases, something that would not have been possible without a much larger team.”
11 points by patethegreat 63 days ago | 2 comments
Augmenting NLQ with language knowledge bases like web search for ChatGPT(hyperarc.com) The rise of warehouses like Snowflake and CDPs like Segment broke down data silos, joining your CRM to your marketing automation, support tickets, and more. This connected view of your business enabled more accurate and actionable insights in traditional BI.
59 points by todsacerdoti 66 days ago | 29 comments
Modern CSV: Multi-Platform CSV File Editor and Viewer(moderncsv.com) Modern CSV is a powerful CSV file editor/viewer application for Windows, Mac, and Linux. Professionals at all levels of technical proficiency use it to analyze data, check files for uploading to databases, modify configuration files, maintain customer lists, and more. We designed it to compensate for the deficiencies of spreadsheet programs in handling CSV/TSV/DSV/etc. files. We strive to create a user experience our customers describe as “blissful”.
23 points by AmeliaWampler 72 days ago | 6 comments
Apache Iceberg now supports geospatial data types natively(wherobots.com) Geospatial solutions were thought of as “special”, because what modernized the data ecosystem of today, left geospatial data mostly behind. This changes today. Thanks to the efforts of the Apache Iceberg and Parquet communities, we are excited to share that both Iceberg and Parquet now support geometry and geography (collectively the GEO) data types.
PostgreSQL Best Practices(speakdatascience.com) PostgreSQL (Postgres) is one of the most powerful and popular relational database management systems available today. Whether you’re a database administrator, developer, or DevOps engineer, following best practices ensures optimal performance, security, and maintainability of your database systems.
72 points by lemonwaterlime 76 days ago | 45 comments
Over 700M events/second: How Cloudflare makes sense of too much data(cloudflare.com) Cloudflare's network provides an enormous array of services to our customers. We collect and deliver associated data to customers in the form of event logs and aggregated analytics. As of December 2024, our data pipeline is ingesting up to 706M events per second generated by Cloudflare's services, and that represents 100x growth since our 2018 data pipeline blog post.
Apple Passwords is hostile to backups(lapcatsoftware.com) In my view, a useful backup system must be (1) chronological, (2) granular, and (3) redundant. A chronological backup system includes multiple historical snapshots of your data, allowing you to recover not only the latest version of your data but also past data that has been deleted or edited. A granular backup system allows you to selectively recover specific fragments of data from your backup without disturbing, deleting, or corrupting the rest of your current data.
Data Branching for Batch Job Systems(isaacjordan.me) Data is being increasingly treated like code has been treated for decades. For many use-cases it isn't enough to know "What is the current value?" but also "What was the value previously?", "Who last changed the value?", and "Why did they change the value?"