A $41,200 humanoid robot was unveiled in China

The Chinese company UBTech Robotics presented a humanoid robot for 299,000 yuan ($41,200). This is reported by SCMP.

Tien Kung Xingzhe was developed in collaboration with the Beijing Humanoid Robot Innovation Center. It is available for pre-order, with deliveries expected in the second quarter.

The robot is 1.7 meters tall and can move at speeds of up to 10 km/h. Tien Kung Xingzhe easily adapts to a variety of surfaces, from slopes and stairs to sand and snow, maintaining smooth movements and ensuring stability in the event of collisions and external interference.

The robot is designed for research tasks that require increased strength and stability. It is powered by the new Huisi Kaiwu system from X-Humanoid. The center was founded in 2023 by UBTech and several organizations, including Xiaomi. He develops products and applications for humanoids.

UBTech’s device is a step towards making humanoid robots cheaper, SCMP notes. Unitree Robotics previously attracted public attention by offering a 1.8-meter version of the H1 for 650,000 yuan ($89,500). These robots performed folk dances during the Lunar New Year broadcast on China Central Television in January.

EngineAI’s PM01 model sells for 88,000 yuan ($12,000), but it is 1.38 meters tall. Another bipedal version, the SA01, sells for $5,400, but without the upper body.

In June 2024, Elon Musk said that Optimus humanoid robots will bring Tesla’s market capitalization to $25 trillion.

Event-Driven Architectures for AI Applications: Patterns & Use Cases

The landscape around Artificial Intelligence (AI) is always changing, which increases the demand for flexible, scalable, and real-time systems. During the development of AI applications, the Event Driven Architecture (EDA) approach enables flexible responsiveness to optimisation needs at a structural level. This note accompanying ExploreStack’s editorial calendar attempts to capture the essence, structure, and patterns as well as cases and other aspects of EDA in relation to AI, with particular focus placed on defining boundaries for technical managers and practitioners.

Exploring Event-Driven Architecture

In comparison to other software constructions, EDA – event-driven architecture – stands out as it allows various applications to respond to events in real time while also enhancing scalability and coupling. An event can be anything that is of importance like a user changing data, activating an element, or changing some sort of system information that states and needs feedback. Unlike the traditional request-and-response architecture, EDA allows for asynchronous communications where individual components can publish and subscribe to independent events that are happening. This is particularly important in AI applications that tend to work with huge quantities and need to process them in a timely manner so inferences and actions can be provided on time.  

The AI application’s relevance mostly comes from the fact that EDA is able to respond to highly responsive data workloads. For instance, AI models may be required to process data on stream, take action regarding predictive actions, or inject themselves with new sets of information cyclically. Because of how EDA is built, with decoupling of components, guarantees flexibility, responsiveness in real time, and the ability to scale, all essential for today’s modern AI systems, makes it ideal.

Key Patterns in AI Event-Driven Applications

Research and industry practices have defined several patterns within Event Driven Architecture (EDA) that are particularly useful for AI applications. These patterns solve certain problems, augment the efficiency of AI systems and improve their effectiveness:

  1. Asynchronous Inference  
    • Most AI models, especially image generation models or those that rely on Large Language Models (LLMs), require a great deal of computation and take an equally lengthy time to complete. This is compounded in synchronous systems where there is little to no user interaction. EDA solves this problem by enabling applications to publish inference requests as events that are taken care of by other components. These components can range from workers to serverless functions, which perform the task and publish results back as events, notifying the application when they’re finished. Such systems are more responsive, use resources better, and can manage much higher levels of concurrency, as witnessed in Stable Diffusion applications where asynchronous inference decreases idle time during peak demand periods.
  2. Real-time Data Updates
    • AI models are only as effective as the data they are trained on, and in many applications, data is dynamic, requiring periodic updates or retraining. Events can trigger these updates automatically when new data arrives or when specific conditions are met, such as a threshold number of new records. This ensures the model remains relevant and accurate over time without manual intervention. For example, in conversational search systems, scheduled tasks and workflows configured via EDA ensure timely and accurate data updates in knowledge bases, leveraging event-driven advantages for enhanced user experience.
  3. Event-Triggered Actions
    • AI can analyse events to detect patterns, anomalies, or predictions and trigger further actions within the system. For instance, user behavior events can lead to personalised recommendations, while fraud detection events can initiate alerts or block transactions. This pattern enables proactive and personalised interactions, enhancing user engagement and system efficiency. It is particularly useful in scenarios where immediate action is required, such as in financial systems where real-time fraud detection is critical.
  4. Decoupling Components
    • Complex AI systems often comprise multiple components, such as data ingestion, preprocessing, model training, and prediction, which need to work together but can be managed independently. EDA facilitates this decoupling by using events as the means of communication, allowing each component to operate separately. This modularity makes it easier to scale, maintain, and update individual parts without affecting the entire system, enhancing overall system resilience and flexibility. This pattern is evident in microservices architectures, where AI components can scale independently based on demand.

Use Cases and Practical Applications

EDA’s application in AI is demonstrated through various use cases, each addressing specific business needs and leveraging the patterns discussed. These use cases highlight how EDA can transform AI applications, improving performance and user experience:

  1. Chatbots and Virtual Assistants
    • In this scenario, user messages are treated as events that trigger natural language processing (NLP) analysis. Based on the intent and entities extracted, further events are generated to fetch data from databases, call external APIs, or perform other actions. Responses are then formatted and sent back to the user via events, enabling efficient handling of concurrent queries and seamless integration with various services. This approach is crucial for maintaining real-time interactions, as seen in AI chatbots that use message queues for efficient information transmission, enhancing user loyalty through proactive, human-like communications.
  2. Recommendation Systems
    • Recommendation systems rely on user interactions, such as clicks, purchases, or ratings, to provide personalized suggestions. These interactions generate events that update user profiles in real-time, triggering the recommendation engine to recalculate and update recommendations. This ensures that suggestions are always based on the latest behavior, enhancing personalization and relevance. For example, e-commerce platforms use EDA to deliver up-to-date product recommendations, improving customer satisfaction and conversion rates.
  3. Fraud Detection
    • In financial institutions, each transaction is an event analyzed by an AI model trained to detect patterns indicative of fraud. If the model identifies a suspicious transaction, it publishes an event to trigger further investigation or block the transaction, enabling real-time detection and response. This use case is critical for reducing financial losses and improving security, with EDA facilitating immediate action based on AI insights.
  4. Predictive Maintenance
    • In IoT applications, sensor data from machinery is streamed as events into the system. These events are processed by an AI model that predicts the likelihood of equipment failure. If the prediction indicates a high risk, an event is published to notify maintenance personnel or automatically schedule maintenance tasks, reducing downtime and optimizing maintenance schedules. This is particularly valuable in manufacturing, where EDA ensures timely interventions based on AI predictions.
  5. Personalised Marketing
    • Customer interactions, such as visiting certain pages or clicking on ads, generate events that build customer profiles. AI models analyze these profiles to determine the most effective marketing messages for each customer. When a customer meets specific criteria, such as not making a purchase in a while, an event triggers the sending of a personalized message, improving engagement and conversion rates. This use case demonstrates how EDA can enhance customer experiences through targeted communications.

An interesting observation is how EDA supports personalised marketing, an unexpected application where customer behaviour events trigger tailored messages, boosting engagement in ways not immediately obvious from traditional AI use cases.

Implementation Considerations

When implementing EDA for AI applications, several key considerations ensure the system’s effectiveness and reliability:

  • Choosing the Right Event Broker: Select a robust event broker capable of handling the volume and variety of events, such as Apache Kafka, RabbitMQ, Amazon EventBridge, or Google Cloud Pub/Sub. The choice depends on factors like scalability, latency, and integration with existing systems.
  • Designing Events and Event Schemas: Define clear and consistent event schemas to ensure all components understand the structure and meaning of the events, including event type, payload, and metadata. This is crucial for maintaining interoperability and avoiding errors in event processing.
  • Handling Failures and Retries: Implement mechanisms to handle event processing failures, such as retries with exponential backoff, dead-letter queues for unprocessed events, or alerting systems for manual intervention. This ensures system resilience, especially in high-volume AI applications.
  • Monitoring and Debugging: Use monitoring tools to track event production, consumption, and processing times, identifying bottlenecks and ensuring system performance. Tools like Application Real-Time Monitoring Service (ARMS) at Alibaba Cloud ARMS can be instrumental for long-term operations and maintenance.
  • Security and Compliance: Ensure the event-driven system adheres to security best practices, such as encryption of event data, access controls, and compliance with relevant regulations like GDPR or HIPAA, to protect sensitive AI data and maintain trust.

Comparative Analysis: Challenges and Solutions

To further illustrate, consider the following table comparing challenges in AI applications and how EDA addresses them, based on insights from industry practices:

To further illustrate, consider the following table comparing challenges in AI applications and how EDA addresses them, based on insights from industry practices:This table, derived from Alibaba Cloud Community, highlights how EDA tackles specific AI challenges, reinforcing its suitability for these applications.

Conclusion and Future Outlook

EDA provides a flexible and scalable framework that is particularly well-suited for AI applications. By leveraging patterns such as asynchronous inference, real-time data updates, event-triggered actions, and component decoupling, organizations can build AI systems that are responsive, efficient, and adaptable to changing requirements. The use cases, from chatbots to predictive maintenance, demonstrate practical applications that enhance business outcomes and user experiences.

Looking forward, as AI continues to advance and integrate more deeply into various aspects of business and society, the importance of robust, event-driven architectures will only grow. Technical leaders, particularly CTOs, can position their organizations at the forefront of this evolution by adopting EDA, delivering innovative and high-impact AI solutions that meet the demands of a dynamic digital landscape.

Unleashing Powerful Analytics: Harnessing Cassandra with Spark

Authored by Abhinav Jain, Senior Software Engineer

The adoption of Apache Cassandra and Apache Spark is a game-changer for organizations seeking to change their analytics capabilities in the modern world driven by data. With its decentralized architecture, Apache Cassandra is highly effective in dealing with huge amounts of data while ensuring low downtime. This occurs across different data centers which can be said as well for both fault tolerance and linear scalability: the reason why more than 1,500 companies — such as Netflix and Apple — deploy Cassandra. On the other hand, Apache Spark further boosts this system by processing data in memory, allowing speeds up to 100 times faster than disk-based systems and greatly enhancing the setup introduced by Cassandra.

A fusion of Cassandra and Spark results in not just a speedup, but an improvement of data analytics quality. The organizations that use this report drastically decrease their data processing time from hours to minutes — vital for finding insights quickly. This has brought them closer to staying ahead in the competitive markets since the two technologies work well together: When used jointly, Spark and Cassandra are best suited for real-time trend analysis.

On top of that, the integration of these two technologies is proposed as a response to the growing demand for flexible and scalable solutions in areas as broad as finance, where integrity, validity and speed play an important role. This coaction helps organizations not only control larger sets more expediently but also find valuable intelligence with a pragmatic approach: the decision is made based on their operation or the strategic move of their business. Given this, it becomes evident that knowledge about Cassandra’s integration with Spark should be part of every organization that intends to improve its operational analytical data.

Preface: Combining Cassandra’s Distribution with Spark’s In-Memory Processing

The use of Apache Cassandra has been a common choice for organizations that have large volumes of data to manage since they need distributed storage and handling capabilities. However, its decentralized architecture and tunable consistency levels — along with the ability to distribute large amounts of data across multiple nodes — is what makes it ideal without introducing minimal delays. In contrast, Apache Spark can work out processing and analyzing data in memory, which complements Cassandra as an outstanding partner able to deliver real-time analytics plus batch processing tasks.

Setting Up the Environment

To optimally prepare the environment for analytics using Cassandra and Spark, you start the process by installing Apache Cassandra first, then launching a Spark cluster. Both components need individual attention during configuration to promote harmony and achieve the best output from each side. The inclusion of connectors like DataStax Spark Cassandra Connector or Apache Spark Cassandra Connector is pivotal, since they help in effective data flow between Spark and Cassandra systems. Such connectors enhance query operation through Spark’s easy access to data from Cassandra without much network overhead due to parallelism optimization.

With the connectors having been configured, it’s equally vital that you tinker with the settings in a bespoke manner to cater to the workload specifics and volume of data. This could entail tweaking Cassandra’s compaction strategies and Spark’s memory management configurations — adjustments that must be made with anticipation of the incoming data load. The last leg of this journey is verifying the setup through test data: the successful integration signals workability, enabling a seamless analytics operation with due expectations. This setup — robust and intricate — acts as a fulcrum for both technologies, allowing them to be used at full capacity in one coherent analytics environment.

Performing Analytics with Spark and Cassandra

A fusion of Spark with Cassandra results in an enhancement of data processing: through the utilization of Spark’s efficient distribution model and Cassandra’s powerful computing capabilities. The end users are therefore able to perform advanced queries and deal with large datasets easily using Cassandara’s direct storage framework. In addition, these capabilities are enhanced by a number of libraries embedded within Spark, such as MLlib for machine learning, GraphX for graph processing, and Spark SQL for structured data handling — tools that support easy execution of complex transformations, and predictive analytics and data aggregation tasks. Furthermore, by caching data in memory, Spark speeds up iterative algorithms and queries, thus making it ideal where frequent data access is needed, coupled with manipulation via an intuitive user interface. The integration improves workflow and maintains high performance even after scaling to meet growing demands on big data across landscapes where large amounts prevail.

Real-time Analytics and Stream Processing

Furthermore, Spark plus Cassandra real-time analytics is a good approach to organizations’ intake and immediate analysis of data flows. This value is especially important for the business where speed and informativity are important. For example, monitoring of financial transactions, social network activity or IoT output information. Through Spark Streaming, data can be ingested in micro-batches and processed continuously with the possibility of implementing complex algorithms on the fly. When Spark is used with the CDC feature from Cassandra or tightly integrated with Apache Kafka as part of message queuing infrastructure, it turns into a powerful weapon that allows development teams to craft feedback-driven analytical solutions supporting dynamic decision processes which adapt towards changes unearthed from incoming data streams.

Machine Learning and Advanced Analytics

In addition to traditional analytics tasks, Spark opens up possibilities for advanced analytics and machine learning with Cassandra data. Users can create and model machine learning from Cassandra-stored data without having to move or duplicate it, hence enabling predictive analytics and anomaly detection as well as other high-end use cases through the adoption of Spark’s MLlib plus ML packages.

Best Practices and Considerations

One must take into account the best practices when integrating Spark and Cassandra for advanced analytics so that their potential can be maximized effectively. To ensure this, it is important to modify the data model of Cassandra in a way that meets the query patterns, helping reduce read and write latencies. In addition, when using partition keys design, distribute data equally across nodes to prevent hotspots while also configuring Spark’s memory and core settings appropriately. This will help you avoid resource overcommitment and thus any unnecessary performance issues.

Moreover, monitoring of both Spark and Cassandra clusters should be maintained continuously. Make use of tools such as Apache Spark’s web UI and Cassandra’s nodetool that can help you with performance metrics which would lead to bottlenecks showing up in no time. You must put in place strict data governance policies; this involves carrying out regular audits and compliance checks, which would ensure data integrity and security. Ensure secure access to data using authentication plus encryption (both in transit and at rest) that prevents unauthorized access and breaches.

Conclusion

Combining Apache Cassandra and Apache Spark creates a significant platform for large-scale analytics: it helps organizations get valuable and meaningful data much quicker than they ever did. By taking advantage of what each technology does best, companies have the opportunity to stay ahead of the competition, foster innovation, and ensure their decisions are based on quality data. Be it historical data analysis, streaming data processing as it flows or constructing machine learning frameworks, Cassandra and Spark, when brought together, form an adaptable and expandable solution for all your analytical needs. 

Data Fabric and Data Mesh: Complementary Forces or Competing Paradigms?

As the world continues to change, two frameworks have emerged to help businesses each manage their data ecosystems – Data Fabric and Data Mesh. While both these frameworks aim to simplify a business’s data governance, integration, and access, they differ quite a lot in their philosophy and how they operate. Data Fabric focuses more on technological orchestration over a distributed environment. Alternatively, Data Mesh focuses more on structural decentralization and domain-centric autonomy. This article looks at the powerful cloud-based architecture that integrates these two frameworks through its definitions, strengths, limitations, and the potential for synergy.

What is Data Fabric?

The Data Fabric concept originated in 2015 and came into focus after Gartner included it in the top analysis trends of 2020. In the DAMA DMBOK2 glossary, data architecture is defined as the plan for how to manage an organization’s data assets in a way that model of the organization’s data structures. Data Fabric implements this by offering a unified framework that automatically and logically integrates multiple disjointed data systems into one entity. 

Simply put, Data Fabric is a singular architectural layer that sits on top of multiple heterogeneous data ecosystems – on-premises systems, cloud infrastructures, edge servers –  and abstracts their individual complexities. It uses and combines several data integration approaches like the use of special data access interfaces (APIs), reusable data pipelines, automation through metadata, and AI orchestration to provide and facilitate non-restricted access and processing. Unlike older methods of data virtualization, which assisted in constructing a logical view, Data Fabric combines with the essence of containerization, which allows better management, control, and governance making masking it more powerful for modernizing applications than traditional methods.

Key Features of Data Fabric

  • Centralized Integration Layer: A virtualized access layer unifies data silos, governed by a central authority enforcing enterprise standards.
  • Hybrid Multi-Cloud Support: Consistent data management across diverse environments, ensuring visibility, security, and analytics readiness.
  • Low-Code/No-Code Enablement: Platforms like the Arenadata Enterprise Data Platform or Cloudera Data Platform simplify implementation with user-friendly tools and prebuilt services.

Practical Example: Fraud Detection with Data Fabric

Consider a financial institution building a fraud detection system:

  1. An ETL pipeline extracts customer claims data from multiple sources (e.g., CRM, transaction logs).
  2. Data is centralized in a governed repository (e.g., a data lake on Hadoop or AWS S3).
  3. An API layer, enriched with business rules (e.g., anomaly detection logic), connects tables and exposes the unified dataset to downstream applications.


While this approach excels at technical integration, it often sidesteps critical organizational aspects – such as data ownership, trust, and governance processes—leading to potential bottlenecks in scalability and adoption.

How Data Mesh Works

Data Mesh, introduced around 2019, is a new framework of data architecture that puts a greater emphasis on people rather than technology and processes. Like DDD, Data Mesh advocates for Domain-oriented decentralization, which promotes the fragmentation of data ownership among business units. Unlike Data Fabric, which controls everything from a single point, Data Mesh assigns domain teams with the responsibility of treating data as a product that can be owned, accessed, and interacted with in a self-service manner. 

Core Principles of Data Mesh

  • Domain-Oriented Decentralization: The closest teams to the data, whether it be its consumption or generation, have the ownership and management of the data. 
  • Data as a Product: More than just a simple dataset, each dataset can be marketed and comes with features such as access controls and metadata. 
  • Self-Service Infrastructure: Centralized domain teams are able to function autonomously because of a centralized platform. 
  • Federated Governance: Domains without a central data governance point are controlled centrally in terms of standards, data policies, and interfacing.

Practical Example: Fraud Detection with Data Mesh

Using the same fraud detection scenario:

  1. A domain team (e.g., the claims processing unit) defines and owns an ETL/ELT job to ingest claims data.
  2.  Datasets (e.g., claims, transactions, customer profiles) are stored separately, each with a designated owner.
  3.  A data product owner aggregates these datasets, writing logic to join them into a cohesive fraud detection model, delivered via an API or event stream.

This approach fosters accountability and trust by embedding governance into the process from the outset. However, its reliance on decentralized teams can strain organizations lacking mature data cultures or robust tooling.

Emerging Tools

Data Mesh is still maturing technologically. Google’s BigLake, launched in 2022, exemplifies an early attempt to support Data Mesh principles by enabling domain-specific data lakes with unified governance across structured and unstructured data.

Data Fabric works best with complex siloed infrastructures since it offers a top-down approach to data access. On the other hand, Data Mesh performs well in decentralized organizations that are willing to undergo a cultural shift and give more emphasis on trust and agility as compared to technical standardization.

Just like data fabric and data mesh, enterprise operational context and digital transformation journey determines the scope of its existence. The cloud provides a platform where both approaches can be integrated. Consider an architecture where there exists an event bus (for example Apache Kafka), which streams data to many different consumers. The consumers could include AWS S3, which acts as a data lake, and ETL pipelines (AirFlow for batch and NiFi for streaming), which serve to integrate operational and historical data. Add a robust Master Data Management (MDM) layer and analytics will be of good quality. 

This is the integration point where synergy shines: the centralized integration of data fabric sets up the infrastructure and data mesh domain autonomy makes it possible to innovate. A cloud native application platform which enables and controls innovation is the result. Business Intelligence (BI) dashboard is an example, which could be drawing the Mesh IoT dashboard clean data products, while Fabric governs seamless access to data.

A Call to Innovate

Marrying these paradigms isn’t without hurdles. Architects and engineers must grapple with:

  • Migration Complexity: How do you transition on-premises data to the cloud without disruption?
  •  Real-Time vs. Batch: Can the platform balance speed and depth to meet business demands?
  •  Data Quality: How do you embed quality checks into a decentralized model?
  •  Security and Access: What federated security model ensures ease without compromising safety?
  •  Lifecycle Management: How do you govern data from creation to destruction in a hybrid setup?


Moreover, the cloud isn’t a silver bullet. Relational databases often fall short for advanced analytics compared to NoSQL, and data lake security models can hinder experimentation. Siloed data and duplication further complicate scalability, while shifting from centralized to decentralized governance requires a cultural leap.

The Verdict: Together, Not Versus

So, is it Data Fabric versus Data Mesh? These methods have no real conflict; they work hand in hand. Data Fabric shows the threads of a technology metaphor for a superordinate access to information, while Data Mesh gives authority to the operational teams to manage their data. In a cloud-powered ecosystem, they have the potential to revolutionize data management by merging centralization’s productivity with decentralization’s creativity. The challenge that arises is not what to select, but how to combine the multifarious assets into a harmonious orchestra that nurtures trust, economic agility, and value to the enterprise. As the instruments undergo changes and institutions transform, these two concepts may as well be the paradigm shift that data architecture has long been waiting for, shaken, stirred and beautifully blended.