Latest Gemini and Nano Banana Enhancements in LangChain4j
A few days ago, LangChain4j 1.11.0 was released, and with this version, a few notable enhancements to the support of the Gemini model family have landed. Let’s dive in!
New Image Generation Models (Gemini 2.5 & 3.0 Preview, aka 🍌 Nano Banana)
Before showing some snippets of code, let me give you the link to the full documentation on the new image model: docs.langchain4j.dev/integrations/image-models/gemini
Read more...Researching Topics in the Age of AI β Rock-Solid Webhooks Case Study
Back in 2019, I spent significant time researching Webhooks. In particular, I was interested in best practices, pitfalls, design patterns, and approaches for implementing Webhooks in a reliable, resilient, and effective way.
Everything is distilled in that article: Implementing Webhooks, not as trivial as it may seem
It likely took me a full week to dive deep into this subject, finding sources and experimenting with design patterns myself. But nowadays, AI makes it easier to dive deeper into topics, explore unfamiliar aspects, and share findings with your team.
Read more...A Javelit Frontend for the Deep Research Agent
A month ago, I wrote about building a research assistant in Java, using the Gemini Interactions API, and the Deep Research agent. Today I wanted to revisit this project, but with the goal to make it more user-friendly, with a Javelit based frontend.
The Research Pipeline: From Query to Infographic
- First, the user enters the subject of the research.
- A button action triggers the research of possible topics about that subject (ie. the different possible facets or angles of the subject), using Gemini 3 Flash with Google Search activated.
- The user selects the facets they’re interested in, to restrict the research to only those aspects.
- Then the Deep Research agent kicks in, via the Gemini Interactions API, and spends a few minutes researching the topic.
- Once the final report is ready, Gemini 3 Pro creates a solid summary.
- The summary is used to generate a sketchnote with Nano Banana Pro.
A Look at the User Interface
Let’s say, as a user, I want to research information about the OpenClaw / MoltBot / ClawdBot personal AI assistant (unless it has again changed its name? 😃)
Read more...Executable Markdown Files with gcli-mdrun & Gemini CLI
Have you ever wanted to turn your cool LLM prompts & tools, research notes, automation ideas, or even a simple “todo” list into an executable script?
Inspired by a HackerNews post about executable Markdown,
I’m happy to share gcli-mdrun, a smart little script that allows you to transform standard Markdown files
into executable scripts powered by Gemini CLI.
This project allows you to create AI-driven automation, pipelines, and autonomous bots using mere Markdown text files. You can find the project on GitHub at https://github.com/glaforge/gcli-mdrun.
Read more...Implementing an arXiv MCP Server with Quarkus in Java
For my recent presentation at SnowCamp on AI Standards & Protocols for AI Agents, I decided to build an MCP server to access the arXiv research paper website where pre-print versions are published and shared with the community.
My goal was to shed light on some lesser-known aspects of the Model Context Protocol:
- 🛠️ While the majority of MCP servers use the tools feature to expose actions that LLMs can request to call,
- 📄 An MCP server can also share resources (and resource templates), exposing various static assets the AI app might be interested in,
- ✏️ And prompts (and prompt templates) that users can access and reuse to utilize the MCP server effectively.
For the impatient, feel free to go straight to the GitHub repository for the full source code. The README.md file gives instructions on how to build, run, configure, and use the server.
Read more...On AI Standards and Protocols: Focus on MCP and A2A
At SnowCamp 2026, with my Cast Codeurs buddy Emmanuel Bernard of Hexactgon, I had the chance to deliver a talk on AI standards and protocols, with a big focus on MCP (Model Context Protocol), and A2A (Agent 2 Agent Protocol).
Without further ado, here’s the slide deck we presented:
This talk is based on the Devoxx 2025 deep dive session that I delivered with Emmanuel and my colleague Mete Atamel. As the talk wasn’t recorded during SnowCamp, I’ll share with you the 3h-long video from Devoxx below:
Read more...Building a Research Assistant with the Interactions API in Java
First of all, dear readers, let me wish you a happy new year! This is my first post on this blog for 2026. I’m looking forward to continuing sharing interesting content with you.
During my holiday break, I wanted to put my recent Java implementation of the Gemini Interactions API to the test. I implemented and released it with the help of Antigravity. My colleague Shubham Saboo and Gargi Gupta wrote a tutorial on how to build an AI research agent with Google Interactions API & Gemini 3. I thought this was a great opportunity to replicate this example in Java using my Interactions API Java SDK.
Read more...Implementing the Interactions API with Antigravity
Google and DeepMind have announced the Interactions API, a new way to interact with Gemini models and agents.
Here are some useful links to learn more about this new API:
- An announcement is available on Google’s Keywords blog:
Interactions API: A unified foundation for models and agents - A more detailed article is available on Google’s developers blog:
Building agents with the ADK and the new Interactions API - The newly released Gemini Deep Research agent is now available via the Interactions API as well:
Build with Gemini Deep Research - The official documentation of the Interactions API.
About the Interactions API
The Rationale and Motivation
The Interactions API was introduced to address a shift in AI development, moving from simple, stateless text generation to more complex, multi-turn agentic workflows. It serves as a dedicated interface for systems that require memory, reasoning, and tool use. It provides a unified interface for both simple LLM calls and more complex agent calls.
Read more...AI Agentic Patterns and Anti-Patterns
This week, I was on stage at the Tech Rocks Summit 2025 in the beautiful ThéÒtre de Paris. This is the first I’m attending this event, gathering a nice crowd of CTOs, tech leads, architects, and decision makers.
My talk focused on what everyone is talking about right now: AI Agents. And in particular, I was interested in sharing with the audience things I’ve seen work or not work in companies, startups, and via tons of discussions with AI practitioners I met at conferences, meetups, or customer meetings.
Read more...Gemini Is Cooking Bananas Under Antigravity
What a wild title, isn’t it? It’s a catchy one, not generated by AI, to illustrate this crazy week of announcements by Google. Of course, there are big highlights like Gemini 3 Pro, Antigravity, or Nano Banana Pro, but not only, and this is the purpose of the article to share with you everything, including links to all the interesting materials about those news.
Gemini 3 Pro
The community was eagerly anticipating the release of Gemini 3. Gemini 3 Pro is a state-of-the-art model, with excellent multimodal capabilities, advanced reasoning, excellent at coding, and other agentic activities.
Read more...