Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 14, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 14, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 14, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 14, 2025

      I test a lot of AI coding tools, and this stunning new OpenAI release just saved me days of work

      May 14, 2025

      How to use your Android phone as a webcam when your laptop’s default won’t cut it

      May 14, 2025

      The 5 most customizable Linux desktop environments – when you want it your way

      May 14, 2025

      Gen AI use at work saps our motivation even as it boosts productivity, new research shows

      May 14, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Strategic Cloud Partner: Key to Business Success, Not Just Tech

      May 14, 2025
      Recent

      Strategic Cloud Partner: Key to Business Success, Not Just Tech

      May 14, 2025

      Perficient’s “What If? So What?” Podcast Wins Gold at the 2025 Hermes Creative Awards

      May 14, 2025

      PIM for Azure Resources

      May 14, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Windows 11 24H2’s Settings now bundles FAQs section to tell you more about your system

      May 14, 2025
      Recent

      Windows 11 24H2’s Settings now bundles FAQs section to tell you more about your system

      May 14, 2025

      You can now share an app/browser window with Copilot Vision to help you with different tasks

      May 14, 2025

      Microsoft will gradually retire SharePoint Alerts over the next two years

      May 14, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Meet Jockey: A Conversational Video Agent Powered by LangGraph and Twelve Labs API

    Meet Jockey: A Conversational Video Agent Powered by LangGraph and Twelve Labs API

    July 6, 2024

    Recent developments in the field of Artificial Intelligence are completely changing the way humans engage with video material. The open-source chat video agent ‘Jockey‘ is a great example of this innovation. Jockey provides improved video processing and interaction by utilizing the potent powers of Twelve Labs APIs and LangGraph. 

    Twelve Labs offers modern video understanding APIs that can extract comprehensive insights from video footage. Its APIs operate directly with video data, analyzing visuals, audio, on-screen text, and temporal correlations, in contrast to traditional methods that rely on pre-generated captions. With this all-encompassing approach, videos are understood more precisely and contextually.

    Classification, question answering, summarization, and video search are some of the main features of Twelve Labs APIs. With the help of these APIs, developers can build apps for various use cases, including AI-generated highlight reels, interactive video FAQs, automated video editing, and content discovery. The scalability and strong enterprise-grade security of these APIs make them ideal for managing large video archives, creating new opportunities for applications that rely on video.

    With the release of LangGraph v0.1 by LangChain, an adaptable framework for creating agentic and multi-agent applications has been presented. With LangGraph’s customizable API for cognitive architectures, developers can more precisely control the flow of code, prompts, and large language model (LLM) calls than they could with LangChain AgentExecutor, its predecessor. Additionally, LangGraph allows for human approval prior to task execution and offers ‘time travel’ capabilities for altering and resuming agent operations, which in turn facilitates human-agent collaboration.

    LangChain introduced LangGraph Cloud, which is presently in closed beta, to supplement this architecture. LangGraph Cloud provides scalable infrastructure for deploying LangGraph agents, and managing servers and task queues to effectively manage several concurrent users and big states. It interfaces with LangGraph Studio and enables real-world interaction patterns to visualize and troubleshoot agent trajectories. Because of this combination, agentic applications can be developed and deployed more quickly.

    With its most recent release, v1.1, Jockey has seen a substantial change compared to its original LangChain-based version. By using LangGraph, Jockey boasts improved scalability and functionality in both frontend and backend operations. This shift has optimized Jockey’s architecture, enabling more accurate and efficient control over intricate video workflows.

    Jockey fundamentally combines the advantages of LLMs with the customizable structure of LangGraph to provide video APIs from Twelve Labs. The complex network of nodes that makes up LangGraph, which includes elements like the Supervisor, planner, video-editing, video-search, and video-text-generation nodes, helps in Jockey’s decision-making. This configuration guarantees smooth execution of video-related operations and quick processing of user requests.

    The fine control LangGraph offers over every stage of the workflow is one of its most notable features. By carefully controlling the information flow between nodes, Jockey can maximize token consumption and improve node response accuracy. Video processing is more successful and efficient as a result of this refined control.

    Jockey’s advanced architecture uses a multi-agent system to manage intricate video-related activities. The Supervisor, Planner, and Workers are the three primary parts of the architecture. As the main coordinator, the Supervisor oversees the process and assigns tasks to other nodes. It manages mistake recovery, ensures the plan is followed and starts replanning when it’s needed.

    The planner is in charge of dissecting intricate user requests into digestible chunks that the Workers can carry out. This part is essential for managing workflows, which include multiple steps in video processing. The Workers carry out activities in accordance with the planner’s strategy and include specialized agents for video search, video text generation, and video editing.

    Jockey’s modular architecture makes extension and customization easier. To accommodate more complicated scenarios, developers can expand the state, change the prompts, or add extra workers for particular use cases. Because of its adaptability, Jockey provides a flexible platform on which to develop sophisticated video AI applications.

    In conclusion, Jockey is a great combination of the advanced video interpretation APIs from Twelve Labs and the adaptable agent framework from LangGraph. This combination creates new opportunities for engagement and intelligent video processing. 

    The post Meet Jockey: A Conversational Video Agent Powered by LangGraph and Twelve Labs API appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleBeyond Deep Learning: Evaluating and Enhancing Model Performance for Tabular Data with XGBoost and Ensembles
    Next Article I Love PDF 2

    Related Posts

    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 15, 2025
    Common Vulnerabilities and Exposures (CVEs)

    CVE-2025-47785 – Emlog SQL Injection and Remote Code Execution

    May 15, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    defineExpose and in Vue 3 for component interaction and theming

    Development

    This smart ring for fitness freaks nearly beats Oura. Here’s how

    Development

    CVE-2025-30667 – Zoom Workplace Apps for Windows Denial of Service

    Common Vulnerabilities and Exposures (CVEs)

    Dragon Age: The Veilguard director has left the studio after 18 years for an opportunity she, “couldn’t turn down” in the RPG space

    News & Updates

    Highlights

    To work with two different drop downs which are dependent on each other

    June 4, 2024

    I have a scenario, which I am finding bit difficult to automate.
    The test scenario is :
    There are two drop down.

    To select the state.
    To select the respective cities.

    So, I need to write a script to automate, where I need to select a state from State drop down and for that state I need to select all the cities one after the other. Once all the cities are selected for that respective state, I need to come out of that loop and select a different state and repeat this process.
    This is the code which i have written.
    def test_state(self):
    element = self.getAllTheElements(self.Register_page_elements,’xpath’, ‘option’)
    for link in element:
    if link.get_attribute(‘value’):
    link.click()
    print(“test”)
    time.sleep(4)
    element = self.getAllTheElements(‘autocomplete-city’, ‘id’, ‘option’)
    for link2 in element:
    if link2.get_attribute(“value”):
    link2.click()
    print(“state”)
    time.sleep(2)

    Since I have not given break statement, it will throw an error after completing entire iteration.
    But, if I give a break it wont work the way I want.
    This is the HTML code of the drop down.
    For state.
    <div class=”under-line”>
    <select type=”text” class=”enque-content” autocomplete=”off” required=”required” id=”autocomplete-state” name=”state” onchange=”getCities(this)”>
    <option selected=”” disabled=”disabled” value=””>Please select your state*</option>
    <option value=”1″>Andaman and Nicobar Island</option><option value=”2″>Andhra Pradesh</option><label alt=”First Name*” class=”enque-placeholder-label” placeholder=”First Name*”> </label><option value=”3″>Arunachal Pradesh</option><option value=”4″>Assam</option><option value=”5″>Bihar</option><option value=”6″>Chandigarh</option><option value=”8″>Dadra and Nagar Haveli</option><option value=”7″>Chhattisgarh</option><option value=”9″>Daman and Diu</option><option value=”10″>Delhi</option><option value=”37″>Goa</option><option value=”12″>Gujarat</option><option value=”13″>Haryana</option><option value=”14″>Himachal Pradesh</option><option value=”15″>Jammu and Kashmir</option><option value=”16″>Jharkhand</option><option value=”17″>Karnataka</option><option value=”18″>Kerala</option><option value=”19″>Lakshadweep</option><option value=”20″>Madhya Pradesh</option><option value=”21″>Maharashtra</option><option value=”22″>Manipur</option><option value=”23″>Meghalaya</option><option value=”24″>Mizoram</option><option value=”25″>Nagaland</option><option value=”26″>Odisha</option><option value=”27″>Puducherry</option><option value=”28″>Punjab</option><option value=”29″>Rajasthan</option><option value=”30″>Sikkim</option><option value=”31″>Tamil Nadu</option><option value=”32″>Telangana</option><option value=”33″>Tripura</option><option value=”34″>Uttar Pradesh</option><option value=”35″>Uttarakhand</option><option value=”36″>West Bengal</option></select>
    </div>

    Once I select the state then the city drop down will appear.

    SolidJS is off to a solid start

    May 23, 2024

    Google to offer free version of Gemini Code Assist

    February 25, 2025

    Google’s Next-Generation AI Model Gemini 1.5 Pro is Now Available in Public Preview on Google Cloud’s Vertex AI Platform

    April 9, 2024
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.