Ahorra un 25 % (o incluso más) en tus costes de Kafka | Acepta el reto del ahorro con Kafka de Confluent
Confluent Private Cloud (CPC) is a new software package that extends Confluent’s cloud-native innovations to your private infrastructure. CPC offers an enhanced broker with up to 10x higher throughput and a new Gateway that provides network isolation and central policy enforcement without client...
Confluent announces the General Availability of Queues for Kafka on Confluent Cloud and Confluent Platform with Apache Kafka 4.2. This production-ready feature brings native queue semantics to Kafka through KIP-932, enabling organizations to consolidate streaming and queuing infrastructure while...
Explore new Confluent Intelligence features: A2A integration, multivariate anomaly detection, vector search for Cosmos DB and S3 Vectors, Private Link, and MCP support.
Confluent Cloud now offers native Kafka Streams health monitoring to simplify troubleshooting. The new UI provides at-a-glance application state, performance ratios to pinpoint bottlenecks (code vs. cluster), and state store metrics.
At Current 2025 in New Orleans (Oct 29–30), developers, data engineers, operators, architects & tech execs unlock real-time data + AI insights.
Confluent is providing our customers and prospects with a full package to build trust and innovate securely with Confluent. With our technical documentation, foundational principles and a new level of transparency.
Powering analytics and AI requires reliable, consistent, and easily discoverable data to reach the data lake. To enforce these needs, strong and holistic governance is an important of building better platforms for getting from raw data to valuable insights and actions.
See how Confluent and its partner ecosystem are making it easier to use real-time data streaming as the fuel for your agentic AI and advanced analytics applications.
Learn how to scale Kafka Streams applications to handle massive throughput with partitioning, scaling strategies, tuning, and monitoring.
Learn how to choose the right Apache Kafka® multi-cluster replication pattern and run an audit-ready disaster recovery and high availability program with lag SLOs, drills, and drift control.
Learn how to handle data transformation, schema evolution, and security in Kafka Connect with best practices for consistency, enrichment, and format conversions.
Confluent Champion blog post featuring Aamir Thoker
Learn best practices for validating your Apache Kafka® disaster recovery and high availability strategies, using techniques like chaos testing, monitoring, and documented recovery playbooks.
Learn best practices for running Kafka Connect in production—covering scaling, security, error handling, and monitoring to build resilient data integration pipelines.
Learn how to automate BI with real-time streaming. Explore event-driven workflows that deliver instant insights and close the gap between data and action.
Discover how banks and payment providers use Apache Kafka® streaming to detect and block fraud in real time. Learn patterns for anomaly detection, risk mitigation, and trusted automation.
Discover why microservices architectures thrive with event-driven design and how streaming powers applications that are agile, resilient, and responsive in real time.