AI
45 TopicsAI Friday Live: NVIDIA GTC - Disney Robots, Blackwell, Groot, Newton And More!
Join us for our first live episode of AI Friday, where we delve into the latest and most exciting developments in the world of artificial intelligence. This week, we discuss the major announcements from NVIDIA's GTC conference, including insights on their new Blackwell chips and the fascinating advancements in autonomous vehicles. Our very own Buu Lam from DevCentral shares his firsthand experience and key takeaways from the event. In addition, we explore the groundbreaking world of robotics, highlighting how Disney's cute robots, powered by NVIDIA and Google DeepMind, are set to revolutionize the industry. We also tease our upcoming episode featuring Steve Wilson, where we will discuss the OWASP Gen AI Security Project and its implications for large language model applications. Don't miss out on this engaging and informative discussion! Would you like some fish with all those new chips? DGX Spark / Station Robotics: Groot and Newton Nemotron Models15Views0likes0CommentsExploring Quantum Computing, AI, Networking and Cryptography at AppWorld 2025
Join us as we dive into the exciting world of quantum computing and AI at App World 2025, held at the fabulous Fontainebleau in Las Vegas. I get to host a fascinating discussion with Daniela Pontes and Brett Wolmarans bwolmarans from F5. We explore the latest advancements in AI, the impact of quantum computing on cybersecurity, and what the future holds for post-quantum cryptography. Discover how F5 is leveraging AI to optimize and secure applications, and learn about the recent release of the AI gateway. Daniela delves into the looming threat of quantum computing on current cryptography standards, explaining the importance of transitioning to quantum-resistant algorithms and even a little on quantum networking. Stay tuned for info on how industries like finance and healthcare are preparing for a quantum future. Don't miss this episode full of expert knowledge and cutting-edge technology!15Views0likes0CommentsAI Friday: Agentic AI, Helix VLAM, Google Gemini Robotics, Managing Agents
Join us for another exciting episode of AI Friday! This week, we delve into the latest advancements and news in the world of artificial intelligence. Our hosts, Aubrey, Lori, Ken, and Byron, discuss the fascinating developments in agentic AI, exploring its implications for autonomy, memory, and perception. We also touch on the potential risks and security concerns associated with this cutting-edge technology. Don't forget to join us on March 28th for a special live stream with Steve Wilson, the founder of the OWASP Top Ten for Large Language Model Applications. It's going to be an insightful session you won't want to miss! Stay tuned for engaging discussions, intriguing insights, and a sneak peek into the future of AI technology. Make sure to hit the like button, subscribe to our channel, and click the notification bell so you never miss an episode of AI Friday! Resources: Rise Of The Thinking Machines Real Robots Gemini Robotics Managing AI Agents50Views0likes0CommentsAlexa+, GPT-5 Pricing, and Livestream Announcement - AI Friday Ep.10
Join us for an exciting AI Friday episode where we dive deep into the latest developments in artificial intelligence! This week, we're bringing you the most pertinent AI news, discussing the intriguing case of psychopathic AI models, and announcing our very first live stream on March 28th with the esteemed Steve Wilson, CTO of Exabeam and a key figure in the OWASP Top Ten for LLM Apps and Generative AI. In this episode, we also explore the launch of Amazon's Alexa+ and its implications for the AI ecosystem. Additionally, we break down the new GPT-5 pricing model, examining its potential impact on the market. Special thanks to Kid Rock for unexpectedly boosting our visibility, plus, get ready to meet our expert guests Ken Arora, Chuck Herrin, Laurie MacVittie, and Joel Moses as they provide their unique insights into the world of AI. Tune in now to stay ahead of the curve in the ever-evolving field of artificial intelligence! Here are the articles we chatted on: Trained Model Goes Psychopath OpenAI Pay-For-Brains Plan Introducing Alexa+ Infinite Retrieval22Views0likes0CommentsSecure AI RAG using F5 Distributed Cloud in Red Hat OpenShift AI and NetApp ONTAP Environment
Introduction Retrieval Augmented Generation (RAG) is a powerful technique that allows Large Language Models (LLMs) to access information beyond their training data. The “R” in RAG refers to the data retrieval process, where the system retrieves relevant information from an external knowledge base based on the input query. Next, the “A” in RAG represents the augmentation of context enrichment, as the system combines the retrieved relevant information and the input query to create a more comprehensive prompt for the LLM. Lastly, the “G” in RAG stands for response generation, where the LLM generates a response with a more contextually accurate output based on the augmented prompt as a result. RAG is becoming increasingly popular in enterprise AI applications due to its ability to provide more accurate and contextually relevant responses to a wide range of queries. However, deploying RAG can introduce complexity due to its components being located in different environments. For instance, the datastore or corpus, which is a collection of data, is typically on-premise for enhanced control over data access and management due to data security, governance, and compliance with regulations within the enterprise. Meanwhile, inference services are often deployed in the cloud for their scalability and cost-effectiveness. In this article, we will discuss how F5 Distributed Cloud can simplify the complexity and securely connect all RAG components seamlessly for enterprise RAG-enabled AI applications deployments. Specifically, we will focus on Network Connect, App Connect, and Web App & API Protection. We will demonstrate how these F5 Distributed Cloud features can be leveraged to secure RAG in collaboration with Red Hat OpenShift AI and NetApp ONTAP. Example Topology F5 Distributed Cloud Network Connect F5 Distributed Cloud Network Connect enables seamless and secure network connectivity across hybrid and multicloud environments. By deploying F5 Distributed Cloud Customer Edge (CE) at site, it allows us to easily establish encrypted site-to-site connectivity across on-premises, multi-cloud, and edge environment. Jensen Huang, CEO of NVIDIA, has said that "Nearly half of the files in the world are stored on-prem on NetApp.”. In our example, enterprise data stores are deployed on NetApp ONTAP in a data center in Seattle managed by organization B (Segment-B: s-gorman-production-segment), while RAG services, including embedding Large Language Model (LLM) and vector database, is deployed on-premise on a Red Hat OpenShift cluster in a data center in California managed by Organization A (Segment-A: jy-ocp). By leveraging F5 Distributed Cloud Network Connect, we can quickly and easily establish a secure connection for seamless and efficient data transfer from the enterprise data stores to RAG services between these two segments only: F5 Distributed Cloud CE can be deployed as a virtual machine (VM) or as a pod on a Red Hat OpenShift cluster. In California, we deploy the CE as a VM using Red Hat OpenShift Virtualization — click here to find out more on Deploying F5 Distributed Cloud Customer Edge in Red Hat OpenShift Virtualization: Segment-A: jy-ocp on CE in California and Segment-B: s-gorman-production-segment on CE in Seattle: Simply and securely connect Segment-A: jy-ocp and Segment-B: s-gorman-production-segment only, using Segment Connector: NetApp ONTAP in Seattle has a LUN named “tbd-RAG”, which serves as the enterprise data store in our demo setup and contains a collection of data. After these two data centers are connected using F5 XC Network Connect, a secure encrypted end-to-end connection is established between them. In our example, “test-ai-tbd” is in the data center in California where it hosts the RAG services, including embedding Large Language Model (LLM) and vector database, and it can now successfully connect to the enterprise data stores on NetApp ONTAP in the data center in Seattle: F5 Distributed Cloud App Connect F5 Distributed Cloud App Connect securely connects and delivers distributed applications and services across hybrid and multicloud environments. By utilizing F5 Distributed Cloud App Connect, we can direct the inference traffic through F5 Distributed Cloud's security layers to safeguard our inference endpoints. Red Hat OpenShift on Amazon Web Services (ROSA) is a fully managed service that allows users to develop, run, and scale applications in a native AWS environment. We can host our inference service on ROSA so that we can leverage the scalability, cost-effectiveness, and numerous benefits of AWS’s managed infrastructure services. For instance, we can host our inference service on ROSA by deploying Ollama with multiple AI/ML models: Or, we can enable Model Serving on Red Hat OpenShift AI (RHOAI). Red Hat OpenShift AI (RHOAI) is a flexible and scalable AI/ML platform builds on the capabilities of Red Hat OpenShift that facilitates collaboration among data scientists, engineers, and app developers. This platform allows them to serve, build, train, deploy, test, and monitor AI/ML models and applications either on-premise or in the cloud, fostering efficient innovation within organizations. In our example, we use Red Hat OpenShift AI (RHOAI) Model Serving on ROSA for our inference service: Once inference service is deployed on ROSA, we can utilize F5 Distributed Cloud to secure our inference endpoint by steering the inference traffic through F5 Distributed Cloud's security layers, which offers an extensive suite of features designed specifically for the security of modern AI/ML inference endpoints. This setup would allow us to scrutinize requests, implement policies for detected threats, and protect sensitive datasets before they reach the inferencing service hosted within ROSA. In our example, we setup a F5 Distributed Cloud HTTP Load Balancer (rhoai-llm-serving.f5-demo.com), and we advertise it to the CE in the datacenter in California only: We now reach our Red Hat OpenShift AI (RHOAI) inference endpoint through F5 Distributed Cloud: F5 Distributed Cloud Web App & API Protection F5 Distributed Cloud Web App & API Protection provides comprehensive sets of security features, and uniform observability and policy enforcement to protect apps and APIs across hybrid and multicloud environments. We utilize F5 Distributed Cloud App Connect to steer the inference traffic through F5 Distributed Cloud to secure our inference endpoint. In our example, we protect our Red Hat OpenShift AI (RHOAI) inference endpoint by rate-limiting the access, so that we can ensure no single client would exhaust the inference service: A "Too Many Requests" is received in the response when a single client repeatedly requests access to the inference service at a rate higher than the configured threshold: This is just one of the many security features to protect our inference service. Click here to find out more on Securing Model Serving in Red Hat OpenShift AI (on ROSA) with F5 Distributed Cloud API Security. Demonstration In a real-world scenario, the front-end application could be hosted on the cloud, or hosted at the edge, or served through F5 Distributed Cloud, offering flexible alternatives for efficient application delivery based on user preferences and specific needs. To illustrate how all the discussed components work seamlessly together, we simplify our example by deploying Open WebUI as the front-end application on the Red Hat OpenShift cluster in the data center in California, which includes RAG services. While a DPU or GPU could be used for improved performance, our setup utilizes a CPU for inferencing tasks. We connect our app to our enterprise data stores deployed on NetApp ONTAP in the data center in Seattle using F5 Distributed Cloud Network Connect, where we have a copy of "Chapter 1. About the Migration Toolkit for Virtualization" from Red Hat. These documents are processed and saved to the Vector DB: Our embedding Large Language Model (LLM) is Sentence-Transformers/all-MiniLM-L6-v2, and here is our RAG template: Instead of connecting to the inference endpoint on Red Hat OpenShift AI (RHOAI) on ROSA directly, we connect to the F5 Distributed Cloud HTTP Load Balancer (rhoai-llm-serving.f5-demo.com) from F5 Distributed Cloud App Connect: Previously, we asked, "What is MTV?“ and we never received a response related to Red Hat Migration Toolkit for Virtualization: Now, let's try asking the same question again with RAG services enabled: We finally received the response we had anticipated. Next, we use F5 Distributed Cloud Web App & API Protection to safeguard our Red Hat OpenShift AI (RHOAI) inference endpoint on ROSA by rate-limiting the access, thus preventing a single client from exhausting the inference service: As expected, we received "Too Many Requests" in the response on our app upon requesting the inference service at a rate greater than the set threshold: With F5 Distributed Cloud's real-time observability and security analytics from the F5 Distributed Console, we can proactively monitor for potential threats. For example, if necessary, we can block a client from accessing the inference service by adding it to the Blocked Clients List: As expected, this specific client is now unable to access the inference service: Summary Deploying and securing RAG for enterprise RAG-enabled AI applications in a multi-vendor, hybrid, and multi-cloud environment can present complex challenges. In collaboration with Red Hat OpenShift AI (RHOAI) and NetApp ONTAP, F5 Distributed Cloud provides an effortless solution that secures RAG components seamlessly for enterprise RAG-enabled AI applications.280Views1like0CommentsA Deep Dive Into AI Villainy - AppWorld 2025
Hey, Community! If you don't know me, I'm Aubrey King, one of our Community Evangelists here at DevCentral, and 14.5 year veteran at F5. I'm also a core and founding member of the OWASP Top 10 For LLM Applications, a project that's had considerable growth and media attention over the past 2 years. Between OWASP and DevCentral, you've probably seen one of my videos or podcasts! Those who have already know that I love to include LOTS of pop culture references, sci-fi, and fantasy to get my message out. Next week at AppWorld 2025, I'll be presenting with Jay Kelley in a split presentation where I talk about AI security threats, and he follows up with how F5 can address these threats. The presentation is a secret and I don’t want to spoil the surprise for people, but I can tell you that I do a comparison of a couple classic AI villains and talk about which one is superior in design, and why. How have these villains influenced the minds of AI developers? For two fictional characters that are over 50 years old, which author had a better idea of how we build AI today? AppWorld 2025 attendees will learn: A better understanding of AI application architectures How to scale AI effectively and efficiently How to incorporate security in app architecture If you’re in management, dev, or infrastructure, this talk will be right up your alley. If AI is on your radar right now, this session is for you!37Views0likes0CommentsThe App Delivery Fabric with Secure Multicloud Networking
This tutorial with accompanying workflow guide deploys customer edge sites and uses Distributed Cloud Multicloud Networking App Connect to establish a Secure MCN App Delivery Fabric, enabling only Layer7 app connectivity between two cloud sites. Manual and automation workflows show how to make this NetOps and DevOps task come to life.300Views2likes0CommentsUsing Distributed Application Security Policies in Secure Multicloud Networking Customer Edge Sites
This tutorial and workflow guide deploys and uses F5 Distributed Cloud App Security policies with apps at local customer edge sites. Deploy a policy in any customer edge site regardless of location in the cloud or on-prem. Manual and automation workflows show how to make this NetOps and DevOps friendly solution come to life.369Views0likes0CommentsAI Friday: OpenAI Roadmap, Weaponized Agentic AI
Welcome to another exciting episode of AI Friday! This week, we dive deep into the latest developments in AI technology and news. Join us as we discuss key highlights from the OpenAI roadmap, including Sam Altman's recent announcements on GPT-4.5, GPT-5 and the future of reasoning models. We also explore the impact of DeepSeek's advancements on the AI landscape and the intriguing bid by Elon Musk to acquire OpenAI's nonprofit governance arm. Before we wrap up, we'll share some important updates about AppWorld in Las Vegas happening next week. If you're interested in application delivery and AI advancements, this is an event you won't want to miss. Don't forget to like, subscribe, and click the link in the comments for more information about AppWorld. Stay tuned for more insightful discussions and exciting announcements! OpenAI Roadmap - https://x.com/sama/status/1889755723078443244 Musk OpenAI Bid - https://www.reuters.com/technology/artificial-intelligence/musk-withdraw-openai-bid-if-board-agrees-charity-terms-lawyers-say-2025-02-13/ Weaponized Agents - https://www.securityweek.com/how-agentic-ai-will-be-weaponized-for-social-engineering-attacks/31Views0likes0Comments