OpenProject 17.3: Evolving backlogs and sprints

OpenProject 17.3: Evolving backlogs and sprints

Video by OpenProject | Open Source Project Management via YouTube
OpenProject 17.3: Evolving backlogs and sprints

The release brings various features and improvements for you, e.g.

00:00 – Introduction
0:15 – A major update: Evolving backlogs and sprints
1:34 – Action boards released to community
2:06 – Improved project home page
2:24 – Sharing of meeting templates (Basic plan and higher)
2:48 – Option to safely change project identifiers
3:03 – Improved workflow configuration

Find out more about all features and improvements in our release notes: https://www.openproject.org/docs/release-notes/17-3-0/

Source

AI For The Open Source Community

AI For The Open Source Community

Video by Open Source Connect via YouTube
AI For The Open Source Community

๐Ÿ—“ Date: 3rd March
โฐ Time: 8:30 PM IST
๐ŸŽค Speaker: Sebastiano Fuccio

Donโ€™t forget to like, share, and subscribe for more talks on AI and emerging technologies!

#OpenSource #ai #cloud #TechTalk #Developers #SebastianoFuccio

Source

Are Kenyan Developers Actually Making Money? (Some Said NO ๐Ÿ˜ณ)

Are Kenyan Developers Actually Making Money? (Some Said NO ๐Ÿ˜ณ)

Video by OpenSource via YouTube
Are Kenyan Developers Actually Making Money? (Some Said NO ๐Ÿ˜ณ)

Are Kenyan developers actually making money?
We went to JKUAT to find out the truth, from street opinions in Nairobi to real developers at a tech event. Some said devs are strugglingโ€ฆ others are making serious money
We also ran a 30-minute challenge where teams had to come up with real solutions for small businesses, and the results were crazy
Plus:
A senior developer breaks down AI in tech
Insights from the Angular community in Kenya
Raw, unfiltered opinions from devs on the ground
This isnโ€™t just a vlogโ€ฆ itโ€™s the reality of tech in Kenya ๐Ÿ‡ฐ๐Ÿ‡ช
Be honestโ€ฆ do you think developers in Kenya are actually making money?

CHAPTERS
0:00 Highlights
0:50 Nairobi Street Interviews
2:36 Breakfast
2:52 JKUAT
3:20 Tech Event
3:39 Senior Dev Interview
6:00 Event Ends
6:32 30 Mins Challenge
6:47 Brainstorming
8:21 Teams Presenting Ideas
23:35 Winner
23:40 Interview with Angular Lead

Source

Enterprise Prompt Engineering & LLM Testing via h2oGPTe | Part 12

Enterprise Prompt Engineering & LLM Testing via h2oGPTe | Part 12

Video by H2O.ai via YouTube
Enterprise Prompt Engineering & LLM Testing via h2oGPTe | Part 12

How Enterprise h2oGPTe manages prompt templates, version control, and multilingual AI agent deployment at scale.

Bridging predictive models and end users requires well-engineered, maintainable prompts. h2oGPTe provides a centralized prompt library where teams can create, clone, version, and share templates across the organization. The H2O Super Agent connects natural language prompts directly to predictive scoring APIsโ€”enabling real-world actions like addressing customer churn. Multilingual template support and UI localization allow consistent AI behavior to be deployed across global markets.

Technical Capabilities & Resources

โžค Prompt Templates & Libraries: Create, clone, and share prompt templates from a managed organizational catalog.
๐Ÿ”— https://docs.h2o.ai/enterprise-h2ogpte/guide/prompts

โžค Prompt Version Control & Iteration: Define system behaviors, iterate on prompt designs, and manage template settings.
๐Ÿ”— https://docs.h2o.ai/enterprise-h2ogpte/guide/prompts#create-a-prompt-template

โžค Template Sharing Across Teams: Distribute prompt templates for consistent AI behavior organization-wide.
๐Ÿ”— https://docs.h2o.ai/enterprise-h2ogpte/guide/prompts#share-a-prompt-template

โžค Custom Multilingual Prompts: Configure language-specific templates for consistent, localized global AI deployment.
๐Ÿ”— https://docs.h2o.ai/enterprise-h2ogpte/guide/prompts#create-a-prompt-template-for-a-specific-language

Source

Advanced MLflow Tracing: Framework Integrations with LangChain, LlamaIndex, LangGraph (Notebook 1.6)

Advanced MLflow Tracing: Framework Integrations with LangChain, LlamaIndex, LangGraph (Notebook 1.6)

Video by MLflow via YouTube
Advanced MLflow Tracing: Framework Integrations with LangChain, LlamaIndex, LangGraph (Notebook 1.6)

In this sixth episode of this series, Jules Damji dives deep into MLflow’s extensive framework integrations. MLflow supports over 30 different open-source agent-building frameworks, allowing you to automatically trace and evaluate complex AI workflows regardless of your chosen architecture.

This tutorial provides a hands-on comparison of three open source agent building frameworks and demonstrates how MLflow provides full visibility into their execution:
๐Ÿ”น ๐—Ÿ๐—ฎ๐—ป๐—ด๐—–๐—ต๐—ฎ๐—ถ๐—ป: Learn how to use high-level primitives like ChatPromptTemplate and StringOutputParser to build sequential workflows. We demonstrate both simple chains and complex multi-step sequences connected via the pipeline operator.
๐Ÿ”น ๐—Ÿ๐—น๐—ฎ๐—บ๐—ฎ๐—œ๐—ป๐—ฑ๐—ฒ๐˜…: See how to build a Retrieval-Augmented Generation (RAG) system. We walk through creating an in-memory vector index, generating embeddings with OpenAI, and using a query engine to retrieve document-based answers, all while capturing the entire operation trace in MLflow.
๐Ÿ”น ๐—Ÿ๐—ฎ๐—ป๐—ด๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต: For more advanced use cases, we explore building stateful, hierarchical agent workflows. We demonstrate a customer service triage system that uses a supervisor node to classify queries and route them to specialized handlers for billing, tech support, or general inquiries.

Key Takeaways:
๐Ÿ”น ๐—”๐˜‚๐˜๐—ผ๐—บ๐—ฎ๐˜๐—ถ๐—ฐ ๐—ง๐—ฟ๐—ฎ๐—ฐ๐—ถ๐—ป๐—ด: All frameworks integrated with MLflow are automatically traced, capturing inputs, outputs, and intermediate steps without manual instrumentation.8
๐Ÿ”น ๐—™๐—ฟ๐—ฎ๐—บ๐—ฒ๐˜„๐—ผ๐—ฟ๐—ธ ๐—ฆ๐—ฒ๐—น๐—ฒ๐—ฐ๐˜๐—ถ๐—ผ๐—ป: Choose LangChain for sequential chains, LlamaIndex for heavy document indexing, and LangGraph for complex, stateful branching or looping workflows.
๐Ÿ”น ๐—ฉ๐—ถ๐˜€๐—ถ๐—ฏ๐—ถ๐—น๐—ถ๐˜๐˜†: Use the MLflow UI to inspect timelines, verify embeddings, and debug the internal logic of your AI agents.

Resources:
๐Ÿ”— Notebook 1.5: https://github.com/dmatrix/mlflow-genai-tutorials/blob/main/06_framework_integrations.ipynb
๐ŸŽฅ Full Series Playlist: https://youtube.com/playlist?list=PLaoPu6xpLk9EI99TuOjSgy-UuDWowJ_mR&si=jdbAbxTCRuxFxfnG

Source

Visual Anomaly & Novelty Detection Workshop VAND 4.0

Visual Anomaly & Novelty Detection Workshop VAND 4.0

Video by OpenCV via YouTube
Visual Anomaly & Novelty Detection Workshop VAND 4.0

Join our Patreon to support the show: https://patreon.com/opencv

We welcome back the team behind the VAND anomaly-detection challenge, a staple of recent CVPR conferences. VAND brings together cutting-edge research on detecting what doesnโ€™t belong in visual dataโ€”spanning anomaly, novelty, and out-of-distribution detection. Building on three successful editions, VAND 4.0 unites supervised, semi-, and unsupervised approaches, including few-, one-, and zero-shot learning, with a strong focus on real-world impact.

Official site: https://sites.google.com/view/vand4-cvpr2026

Info for CVPR attendees: June 4th (1pm-6pm), 2026 in Denver, CO, USA (In Person) + Zoom (Virtual), Half Day
Room: 601, Posters: Exhibit Hall A

OpenCV is a 501(c)(3) registered non-profit in the United States. See how you can support open source CV & AI: http://opencv.org/support/

Watch along for your chance to win during our live trivia segment, and participate in the live Q&A session with questions from you in the audience.

Become a paid member of the channel to help us make more episodes https://www.youtube.com/channel/UCkrcW82Y2kbgU-U9RaYfgxw/join

Got a cool project of your own? Send it to us and you may be featured https://www.jotform.com/form/233105358823151

Source

Why WideEP Inference Needs Data-Parallel-Aware Scheduling – Maroon Ayoub & Tyler Michael Smith

Why WideEP Inference Needs Data-Parallel-Aware Scheduling - Maroon Ayoub & Tyler Michael Smith

Video by PyTorch via YouTube
Why WideEP Inference Needs Data-Parallel-Aware Scheduling - Maroon Ayoub & Tyler Michael Smith

Why WideEP Inference Needs Data-Parallel-Aware Scheduling – Maroon Ayoub, IBM; Tyler Michael Smith, Red Hat

WideEPโ€”wide expert parallelism fails not because experts are expensive, but because routing ignores where state already lives. In PyTorch LLM serving with vLLM, WideEP fans tokens across many experts while KV caches accumulate unevenly across data-parallel replicas. When routing is unaware of KV placement and per-replica load, requests land on replicas that cannot reuse cache or make progress efficiently and latency spikes as expert fan-out grows.

The fix is not reshaping expert parallelism, but making routing data-parallel aware using signals vLLM already exposes. In this talk, we show how llm-d extends its router to leverage KV-cache locality and load awareness when routing WideEP flows. Rather than treating replicas as interchangeable, the router prefers replicas with warm KV state and available capacity, aligning routing decisions with vLLMโ€™s execution reality and reducing cache fragmentation.

This session walks through how KV-aware, data-parallel routing changes WideEP inference in practice: which signals matter, how routing behavior evolves, and where the gains come from. Attendees leave with a clear mental model for when KV- and load-aware routing unlocks higher throughput.

Source