• One Cerebral
  • Posts
  • How you Should Think About Interfaces with AI/LLMs

How you Should Think About Interfaces with AI/LLMs

PhD Student in CS at UC Berkeley: Shreya Shankar

Credit and Thanks: 
Based on insights from The TWIML AI Podcast with 
Sam Charrington.

Today’s Podcast Host: Sam Charrington

Title

AI Agents for Data Analysis

Guest

Shreya Shankar

Guest Credentials

Shreya Shankar is a PhD student in computer science at UC Berkeley, focusing on data management for machine learning and AI, advised by Dr. Aditya Parameswaran. Her career includes roles as the first ML engineer at Viaduct, a research intern at Google Brain, and a software engineering intern at Facebook. Shankar holds a B.S. in Computer Science from Stanford University and has published research in top venues like SIGMOD, VLDB, and CIDR.

Podcast Duration

47:54

This Newsletter Read Time

Approx. 5 mins

Brief Summary

Shreya Shankar, a PhD student at UC Berkeley, to discuss her innovative project, doc ETL (Extract, Transform, Load), which aims to enhance data processing pipelines using large language models (LLMs). The conversation delves into the complexities of building reliable systems for data extraction and analysis, emphasizing the importance of human interaction in refining AI outputs. Shankar shares her insights on the challenges of data quality and the evolving landscape of human-computer interaction in the context of AI.

Deep Dive

Shreya Shankar shares her insights on the complexities of AI interface design, particularly in the context of her project, doc ETL. This declarative framework aims to optimize data processing pipelines powered by large language models (LLMs), addressing the significant challenges posed by unstructured data. Shankar highlighted the pressing need for effective human-computer interaction (HCI) as AI systems become increasingly integrated into data management tasks. She emphasized that the current chat-based interfaces often fall short, lacking the intuitive design necessary for users to interact meaningfully with AI models. The challenge lies in creating interfaces that allow users to manipulate data directly, rather than relying solely on dialogue boxes that can feel disconnected from the task at hand.

Doc ETL is designed to tackle the issue of data extraction from complex sources, such as police misconduct reports, which can span thousands of pages and contain unstructured information. Shankar illustrated this with a project funded by the state of California, aimed at identifying patterns of officer misconduct across various police departments. The traditional approach of hiring interns to annotate this data is not only inefficient but also prone to human error. Instead, doc ETL allows users to specify high-level prompts for operations, which the system then breaks down into manageable tasks that can be executed accurately by the LLM. This method not only enhances efficiency but also ensures that the data processing is reliable and scalable.

However, Shankar acknowledged the challenges associated with data connectors in ETL systems. Many organizations struggle to integrate unstructured data into their existing data warehouses, which are typically designed for structured data. Currently, doc ETL operates as a research prototype, primarily using local folders to manage data. As the project evolves, Shankar anticipates the need for more robust data connectors that can seamlessly integrate with various data sources, including document clouds and JSON files.

The user interface for document processing remains a critical area of exploration. Shankar and her team are working on developing a UI that allows users to interactively build data processing pipelines. Currently, users write YAML files to specify operations, but the goal is to create a more intuitive interface that enables users to visualize their data processing tasks effectively. This could involve a spreadsheet-like interface or other innovative designs that facilitate direct manipulation of data.

Model support is another crucial aspect of doc ETL. Initially, the project utilized OpenAI models due to their early support for tool calling, but Shankar emphasized the importance of incorporating open-source models, especially in sensitive domains like healthcare. The flexibility to choose between different models will be essential as the project scales and addresses various data processing tasks.

Data extraction tasks, particularly in the context of police misconduct, require a nuanced understanding of the information being processed. Shankar pointed out that intelligent extraction is vital; for instance, extracting all medications from a medical document necessitates domain knowledge to identify what constitutes a medication. The complexity of these tasks underscores the need for effective prompts and human-in-the-loop (HITL) systems that allow users to refine their queries based on intermediate outputs.

Evaluation in data processing is another area where Shankar sees significant potential for improvement. The effectiveness of doc ETL hinges on the quality of its validation prompts, which guide the evaluation of outputs. Shankar noted that the current benchmarks for LLMs often focus on reasoning tasks, leaving a gap in the evaluation of data processing capabilities. She advocates for the development of specific benchmarks that address the unique challenges of data processing, such as the need for flexibility in task decomposition and the subjective nature of "correct" outputs.

The concept of agents and agentic systems plays a pivotal role in the architecture of doc ETL. Shankar explained that the system comprises multiple agents designed to handle various tasks, each requiring robust fault tolerance mechanisms. The complexity of managing these agents is significant, as every point of failure must be accounted for to ensure the system operates smoothly. Shankar's experience has shown that many developers overlook the importance of fault tolerance during the initial stages of building agentic systems, leading to complications down the line.

Looking ahead, Shankar expressed interest in exploring state-based models and long-context LLMs as potential solutions to the challenges posed by data processing tasks. These models could enhance the system's ability to maintain context over extended interactions, which is crucial for tasks that require deep reasoning and understanding of complex data. As the field of AI continues to evolve, Shankar's work on doc ETL and her insights into the intricacies of AI interface design will undoubtedly contribute to the development of more effective and user-friendly data processing systems.

Key Takeaways

  • doc ETL framework aims to streamline data processing using LLMs, addressing the challenges of unstructured data.

  • The importance of human oversight in AI systems is highlighted, particularly in refining prompts and outputs based on intermediate results.

  • The need for innovative user interfaces that enhance interaction with AI models, moving beyond traditional chat formats.

Actionable Insights

  • Organizations dealing with unstructured data should consider implementing LLM-powered frameworks like doc ETL to improve data processing efficiency.

  • Encourage teams to engage in iterative feedback loops when working with AI outputs, allowing for prompt adjustments based on real-time results.

  • Invest in developing user interfaces that facilitate direct manipulation of data, enhancing the user experience and improving the accuracy of AI interactions.

  • Prioritize fault tolerance in AI systems by designing agents that can handle failures without compromising the overall workflow.

Why it’s Important

The insights shared in this podcast are crucial as they address the growing need for effective data management solutions in an era where data is increasingly unstructured and complex. As organizations strive to leverage AI for data analysis, understanding the interplay between human oversight and machine learning becomes vital for ensuring accuracy and reliability. The discussion underscores the importance of developing systems that not only automate processes but also empower users to engage meaningfully with the technology.

What it Means for Thought Leaders

For thought leaders in the AI and data management fields, the podcast serves as a call to action to rethink traditional approaches to data processing. It emphasizes the necessity of integrating human expertise into AI workflows, particularly in sensitive areas like law enforcement. The conversation encourages leaders to explore innovative solutions that enhance user interaction with AI, ultimately leading to more reliable and effective data analysis.

Mind Map

Key Quote

"The more humans iterate, the more it diverges from what the system actually would do if you took them out of the loop."

As the demand for data-driven insights continues to rise, we can expect a shift towards more sophisticated AI systems that prioritize human interaction and fault tolerance. The development of user-friendly interfaces that allow for direct manipulation of data will likely become a standard expectation in AI applications. Additionally, as organizations recognize the limitations of traditional data processing methods, there will be an increased focus on frameworks like doc ETL that leverage LLMs to handle complex data tasks efficiently. This trend will not only enhance data quality but also foster greater trust in AI systems across various industries.

Check out the podcast here:

Latest in AI

1. A groundbreaking World Bank-backed pilot program in Nigeria demonstrated that AI tutoring, specifically using GPT-4, can help students achieve learning gains equivalent to two years of traditional education in just six weeks, with students significantly outperforming their peers in English, digital literacy, and AI skills.

2. Microsoft Research has unveiled MatterGen, an open-source AI model that can generate novel inorganic materials with specific desired properties, potentially revolutionizing the field of materials science. Using a diffusion-based architecture trained on over 600,000 stable materials, MatterGen can create crystalline structures across the periodic table, combining elements and refining atomic structures at a speed and precision far exceeding traditional methods.

3. Meta's SeamlessM4T represents a groundbreaking leap in translation technology, moving beyond traditional multi-step translation processes by enabling direct speech-to-speech translation across 101 input languages and 36 output languages with unprecedented accuracy. The model, which uses a novel parallel data mining technique to learn translations, achieves 23% more accuracy than existing models.

Useful AI Tools

1. Microsoft 365 Copilot Chat: Free AI business assistant with pay-as-you-go agent features.

2. Persana AI: AI-powered sales and marketing agents for pipeline generation.

3. Ray2: Advanced AI video generator with hyper-realistic motion capabilities.

Startup World

1. U.S. announced a massive private sector initiative called Stargate, involving OpenAI, Oracle, and SoftBank, to invest up to $500 billion over the next four years in AI infrastructure development across the United States. The project aims to build new data centers, create over 100,000 jobs, and maintain U.S. leadership in AI technology. An initial investment of $100 billion is set to begin immediately, with technology supplied by Arm, NVIDIA, and Microsoft.

2. AI startup funding in the U.S. reached a record $97 billion in 2024, with venture capitalists increasingly focusing on AI-related investments. This figure represents nearly half of the total $209 billion raised by U.S. startups last year, marking the highest proportion of AI-focused deals on record.

3. European quantum computing startup QuantumSphere announced a €75 million Series B funding round led by EQT Ventures. The company, based in Munich, Germany, is developing a new generation of quantum processors that promise to outperform classical supercomputers in specific computational tasks. This significant investment highlights the growing interest in quantum computing technologies and their potential to revolutionize various industries.

Analogy

Shreya Shankar’s work on doc ETL is like crafting a bridge over a chaotic river of unstructured data. Traditional methods struggle with inefficiency and errors, like shaky stepping stones. By designing intuitive interfaces and fault-tolerant systems, she builds sturdy, scalable paths, enabling users to cross seamlessly into the realm of actionable insights.

What did you think of today's email?

Your feedback helps me create better emails for you!

Loved it

It was ok

Terrible

Thanks for reading, have a lovely day!

Jiten-One Cerebral

All summaries are based on publicly available content from podcasts. One Cerebral provides complementary insights and encourages readers to support the original creators by engaging directly with their work; by listening, liking, commenting or subscribing.

Reply

or to participate.