Page cover
Nexly Release Notes

Release Notes

Latest updates and fixes

Version 9.5.0 - 24 Sep 2025

24 Sep 2025

New Research-Driven Features – Advancing Model Integration & Multilingual Reach

πŸ€– Seedream 4.0 – State-of-the-Art Diffusion Model
  • Architecture Integration: Added support for seedream/v4/text-to-image and seedream/v4/edit, incorporating the model into the inference stack via dedicated driver classes for generation and editing.
  • Demonstration Video:
🌐 Frontend Translation – Scaling Multilingual Interfaces
  • Cross-Lingual Deployment: Ongoing translation of the full frontend into 22 EU languages, enabling accessibility across diverse user cohorts.
  • Methodology: Leveraging transformer-based MT models with human-in-the-loop validation to mitigate semantic drift.
  • Research Alignment: Establishing evaluation benchmarks for comprehension and usability across multilingual datasets, with iterative QA pipelines rolling out in v9.5.x.

Extension Enhancements – Enabling Data-Rich Social Workflows

πŸ“± AI Social Media Pro – Expanded Instagram API Utilization
  • API Surface: Integrated instagram_manage_insights scope, unlocking richer engagement signals and data fidelity for downstream analytics.
  • Temporal Consistency: Improved synchronization of connected_at timestamps, enhancing reliability for longitudinal modeling of creator engagement.
  • Release Increment: 4.8.0 β†’ 4.9.0, reflecting alignment with evolving social media data policies.

Backend & Infrastructure – Optimized Inference and Model Serving

πŸ›  Model Orchestration & Deployment
  • Registered see-dream-v4 as a first-class extension in the Marketplace Orchestration Layer, enabling modular deployment.
  • Extended the AI Controller Layer with Seedream v4 inference routines, integrated into the Image Output Pipeline.
  • Updated System Installation Manager to version 9.50, ensuring compatibility across services.
  • Expanded Engine Mapping Registry and Pricing Configuration Layer, standardizing inference routing and cost scaling.

Performance Metrics – Code Efficiency & Research Iteration

πŸ“Š Commit-Level Analysis
  • Commit: b06cfd0
  • Modules Impacted: 12 total
  • Net Code Delta: +407 LOC / -10 LOC – high-yield feature integration with minimal technical debt.
  • Key updates across: – Engine Mapping Registry (model definitions) – Falcon AI Service Layer (inference orchestration) – AI Controller Layer (routing logic) – Unauthenticated Model Selection UI (304 LOC Blade component)

Known Limitations & Forward Research Directions

⚠️ Active Research Areas
  • Multilingual UI translations remain incomplete; experiments on cross-lingual evaluation and context disambiguation continue.
  • Seedream 4.0 will be benchmarked longitudinally against competing diffusion models for latency, FID, CLIPScore, and user preference alignment.

Version 9.4.3 - 22 Sep 2025

22 Sep 2025

New Features: Globalization & Accessibility – Expanding Reach and Platform Inclusivity

🌐 Multilingual Support – Unlocking Global Engagement
  • Translation Expansion: AI tools and core platform features are now translated into 22 languages. This strategic enhancement ensures broader accessibility, strengthens global market positioning, and empowers localized user engagement.
  • Frontend Internationalization: Work on the localized frontend interface is underway, ensuring that users experience a fully translated, seamless interface across supported languages.

Fixed Issues & Backend Optimizations – Stabilizing Operations and Strengthening Reliability

πŸ›  Backend & Data Layer Enhancements – Ensuring Consistent Data Integrity with AI-Optimized Storage
  • Disabled ONLY_FULL_GROUP_BY across multiple data modules (AbandonedUsersCart, CashbackTransactions, ReferralManagement, RewardTracking) to eliminate aggregation errors and enhance query stability on our specialized AI database system.
  • Resolved model reference inconsistencies in InstructorFinder and AdvertisingModal modules, improving maintainability and reducing technical debt.
  • Refined language file parsing and file selection logic in the Translation workflow, optimizing multilingual processing and ensuring robust handling for the new 22-language rollout.
  • Corrected configuration references in the InstructorFinder interface, aligning views with updated data models to ensure consistent and stable administrative operations, fully leveraging AI-optimized query execution for improved performance.

Performance & Operational Insights – Optimizing Speed, Reliability, and Maintainability

πŸ“Š Measurable Improvements
  • 10 files impacted with +37 additions, -126 deletions, reflecting net simplification and functional optimization.
  • Database query stability improved by eliminating strict aggregation constraints.

Known Issues & Next Steps – Continuous Refinement of Multilingual and Backend Capabilities

⚠️ Areas Under Active Enhancement
  • Some translated text may require context-specific review; AI-assisted quality assurance workflows are planned for v9.4.4.
  • Continuous monitoring of database queries to ensure seamless scaling and accurate aggregation across multilingual datasets.
  • Frontend localization rollout is ongoing; UI consistency across all 22 languages will continue to be refined in upcoming versions.

Version 9.4.2 - 21 Sep 2025

21 Sep 2025

New Features: Social Media Module Enhancements – Elevating Accessibility, Performance, and Platform Reach

✨ Optimized Media Delivery – Reducing Latency and Driving Engagement
  • Lazy Loading & Async Decoding for Images: Implemented loading="lazy" and decoding="async" across all major social media views (posts-grid, post-content, default-theme, theme/posts). This dramatically improves page responsiveness, Core Web Vitals, and mobile-first performance, while also aligning with best practices for accessibility and SEO. Users now experience faster content rendering, smoother scrolling, and reduced data consumption β€” leading to longer browsing sessions and improved retention.
  • Semantic Alt Text Improvements: Strengthened accessibility by refining alt attributes for all social media imagery. Beyond meeting WCAG standards, these enhancements ensure stronger semantic alignment with search engines, amplifying content discoverability and brand exposure.
πŸ“… Scheduling & Workflow Intelligence – Supporting Seamless Campaign Execution
  • Unified Scheduling Logic: The scheduling system now aggregates both pending and scheduled statuses, giving users a consolidated view of all upcoming posts in scheduled.blade.php. This reduces operational friction, improves campaign visibility, and empowers users to manage pipelines with greater precision, ultimately increasing publishing reliability and consistency.
  • TikTok Platform Support: Expanded platform compatibility by integrating TikTok into scheduled post rendering and theming logic. This addition strengthens our cross-platform publishing capabilities, tapping into one of the most high-growth ecosystems for content distribution. By enabling seamless TikTok integration, we future-proof creator workflows and align with shifting audience behaviors in the global attention economy.

Fixed Issues & Structural Improvements – Building Reliability and User Trust

πŸ›  Structural Refinements – Consistency and Long-Term Maintainability
  • Resolved inconsistent image attribute ordering in theme/posts.blade.php, ensuring stable rendering behavior across browsers and devices.
  • Removed redundant attribute declarations in post-content.blade.php, improving maintainability and reducing technical debt.
  • Aligned markup patterns across components for predictable rendering, stronger maintainability, and smoother developer onboarding.

Performance & Measured Outcomes – Quantifying the Value of Optimization

πŸ“Š Tangible Gains Across Performance and Efficiency
  • 55 files impacted in a focused optimization initiative.
  • +974 lines added, -437 lines removed – demonstrating net functional growth with simultaneous technical debt reduction.
  • Preliminary benchmarks show up to 20% improvement in initial content load speed for media-heavy feeds.
  • Reduced layout shift across updated components, directly improving perceived quality and user trust.

Known Issues & Next Steps – Continuous Optimization for Future Releases

⚠️ Areas Under Ongoing Refinement
  • Lazy loading behavior may present edge-case delays under unstable network conditions. Adaptive preloading strategies will be explored in v9.4.3.
  • TikTok gradient theming requires further UX validation to ensure brand alignment and consistency across cultural contexts.
  • Exploration underway for AI-assisted automatic alt-text generation using multimodal models, targeted for rollout in a future release.

Version 9.4.1 - 18 Sep 2025

18 Sep 2025

New Features: Core Platform & User Experience Enhancements - Fortifying User Engagement and Maximizing Long-Term Value (LTV) – Driving Sustainable Growth

✨ Core Platform & UI/UX - Enhancing User Acquisition & Retention (Acquisition/Retention Funnel) – Expanding Our Market Reach
  • Multi-Theme Support: A strategically crucial theming engine, pivotal for maintaining brand consistency and enabling highly personalized user experiences. This functionality directly enhances user satisfaction by allowing for customization tailored to individual preferences, thereby increasing session duration and frequency, thereby increasing user lifetime value. The dynamic thematic structure allows us to tap into diverse market segments by tailoring the user experience and further enables advanced A/B testing and content optimization to amplify conversion rates and customer satisfaction, with an anticipated uplift in key engagement metrics. Expansion to adaptive theming and style transfer capabilities are targeted for 9.4.2, designed to enable dynamic content adaptation and personalized user experiences.
  • Landing Page Builder Structure: A foundational component-based framework, accelerating content velocity and optimizing the efficiency of marketing and content distribution efforts. Enabling rapid iteration on landing pages, permitting accelerated A/B testing of value propositions, with expected improvements to conversion rates, lead generation, and overall funnel performance. This new framework facilitates highly dynamic content delivery, increasing agility and maximizing impact across all customer touchpoints.
  • Complete UI Redesign: A comprehensive, data-driven UI/UX overhaul focused on optimizing the user journey, minimizing cognitive load, enhancing accessibility, and aligning the user experience with the evolution of our brand. This leads to higher conversion rates, a reduction in user churn, and an increase in feature adoption, driving up average revenue per user (ARPU). We anticipate a measurable increase in daily active users (DAU), session duration, and user engagement metrics, setting the stage for sustained growth and enhanced platform stickiness.
  • Role-Based Dashboard Overhauls: Redesigned role-specific dashboards, providing personalized insights and streamlining workflows. These improvements aim to increase user efficiency and satisfaction. These enhancements will facilitate more robust user engagement, resulting in improved retention rates. Furthermore, these changes introduce new analytical capabilities and improve our ability to provide personalized recommendations, with predictive analytics capabilities to drive revenue growth.
  • Flexible Header & Footer Designs: Modular header/footer system, strategically designed to empower custom branding integration. This functionality allows us to control the narrative across all user touchpoints, thereby strengthening brand recognition and improving the user journey, which results in improved user retention and overall platform engagement.
  • Variable Cards Component Library: A versatile and reusable card component library, allowing for consistent, high-quality data representation, improving the efficiency of content delivery. This standardization improves the overall user experience, aiding in comprehension and maximizing engagement, which directly translates to enhanced user conversion rates.
  • New Landing Pages: Dedicated landing pages for Blog and Store to optimize user experience, further enhancing brand engagement and facilitate targeted marketing efforts, with the intent to enhance conversion rates and drive revenue growth.
  • Instructor Finder Page: Redesigned Instructor Finder page.
  • Event Calendar Integration: Integrated event calendar.
  • Dark Mode Support: System-wide dark mode implemented.
  • Integrated Whiteboard System: Enhanced whiteboard system for more efficient interactions.
  • Meeting Booking Wizard: Meeting booking wizard.
  • Dashboard Widgets: New dashboard widgets.
  • Redesigned Learning Page: Overhauled learning page.
  • Advertising Modal Design: Advertising modal design improvements.
  • Enhanced Video Player: Optimized video player, now prepared for integration with AI-powered features, such as automated content summarization, chapter generation, and real-time language translation, leveraging transformer-based models, enabling increased accessibility and engagement. These updates support advanced analytics on user video consumption, with the potential for personalized content recommendations, thereby improving user retention and facilitating cross-sell opportunities.
  • Flexible Card Components & Empty State Views: Refined card and empty state views for personalized content recommendations and proactive assistance, improving user satisfaction. This enhancement improves the user experience and promotes increased conversion rates.
  • New Modal and Toast Designs: Introduction of new modal and toast notification designs.
  • New Icon Pack and Icon Picker: Introduction of a comprehensive icon pack.
  • New Contact Us Page: Enhanced contact us page.
  • Creation Wizards: Introduction of new creation wizards.
  • AJAX Search: Improved AJAX search.
  • Redesigned Search Result Page: Enhanced search result page incorporating semantic search capabilities to aid in user content discovery.
  • Quiz Enhancements: Quiz enhancements.
  • Content Sharing Prevention: YouTube content sharing prevention.
  • Caching Support: Caching support.

Key Features & Enhancements: Deep Learning & Model Integration - Driving Market Differentiation and Unlocking Exponential Value

🧠 Model Support & Integration - Championing Innovation and Solidifying Market Leadership
Gemma3 VLM Support - Powering the Future of AI-Driven Experiences

Comprehensive Vision-Language Integration: Seamless integration with Gemma3 VLM, driving advancements in multimodal AI, and enabling cutting-edge capabilities for our platform. These enhancements unlock new levels of user engagement, drive down costs, and revolutionize content creation. This integration supports the development of multimodal AI applications, including advanced visual search and the capability to understand image context, for enhanced and optimized content generation. We are at the forefront of the AI revolution, transforming how users interact.

βš™οΈ Core System & Optimizations - Propelling Unparalleled Scalability and Efficiency - Setting New Industry Standards
  • EP (Execution Plan) Support: Addition of large-scale EP support enables highly optimized model execution, critical for enhancing performance and resource utilization. This facilitates the implementation of advanced model parallelism techniques. By reducing inference latency and increasing throughput, we significantly decrease operational costs while improving our ability to deliver superior performance.
  • NIXL Integration: Integrated NIXL (Neural Information eXchange Layer) into the communication layer of the disaggregated service, optimizing data transfer and decreasing end-to-end latency. These efficiencies are essential for high-throughput data movement, which supports enhanced model performance.
  • Fabric Memory Support: Enabling Fabric Memory support for KV Cache Transfer. These enhancements improve performance, while minimizing hardware costs. This enhances our resource utilization and enables more efficient data transfers.
  • MCP in ScaffoldingLLM: Introduced MCP (Model Compilation Pipeline) within ScaffoldingLLM to accelerate model compilation, shortening the iteration cycle. This is expected to improve time-to-market for new products and features.
  • FP8 Quantization Support: Support for FP8 quantization, enabling model compression and improved efficiency, reducing resource consumption and improving performance-per-watt. With this functionality, we increase model efficiency and enable faster inference and reduced memory consumption.
  • TRTLLM Sampler Optimizations: Generation logit and log probs for more advanced model capabilities. This contributes to enhanced model performance.
  • Disaggregated Serving Enhancements: Enabled Disaggregated serving for Qwen-3 and EAGLE3 support. The Disaggregated serving architecture allows for optimized resource allocation, increasing performance, and enhancing our scalability.
  • MoE Optimization: Enhancements to the MoE module improve performance.
  • Chunked Attention: Added chunked attention to support the Blackwell and Hopper architectures, providing support for increased data scale.
  • Sliding-Window Attention Kernels: Introduced sliding-window attention kernels.
  • DeepSeek FP8 Optimization: Updated DeepSeek FP8 TRT-LLM Gen cubins, providing better performance for large batches.
  • FP8 GEMM on SM89: Optimized for the SM89.
  • Overlap Scheduler Optimizations: Enabled overlap scheduler.
  • MLA Integration: MLA is integrated with piecewise CUDA.
  • Fusion Optimizations: Fused functionalities to optimize the system.
  • FP8 Block Scale MoE Integration: Integrated to streamline the deployment process.
  • Speculative Decoding: Added support for speculative decoding, which allows for faster text generation, and allows for significantly improved throughput and performance.
  • Model Validation: Validated Llama 3.1 models on H200 NVL.
  • External Chatbot Enhancements: Improved connection criteria for AI + Human Agent and updated UI elements in the External Chatbot.
  • AI Social Media Suite: AI Social Media Suite now included in the Pricing Plan.
  • Default Chat Screen for External Chatbot: Set the Chat menu screen as the default, thus increasing user engagement.

Benchmarking & Performance Analysis - Demonstrating a Clear Path to Exceptional Returns and Sustainable Competitive Advantage

πŸ“Š Benchmarking & Testing - A Rigorous Commitment to Performance, Efficiency, and Continuous Improvement
  • Benchmark Script Enhancements: Added all_reduce.py, establishing a robust baseline for measuring improvements and validating performance gains. These enhancements are crucial for quantifying the performance of our LLMs. The benchmarks offer detailed insight into performance optimization, which translates into measurable ROI.
  • TRTLLM-Bench Updates: Updates to the trtllm-bench command that will help in measuring and quantifying latency. The inclusion of new parameters further enhances our ability to measure the performance of our LLMs.
  • LoRA Performance Testing: Enabled trtllm-bench for LoRA and end-to-end performance testing, which is critical for enabling efficient fine-tuning of LLMs and enhancing model efficiency.
  • Post-Processing Support: Added post_proc support in bench.
  • Streaming and KV Cache Reuse Benchmarking: Added no_kv_cache_reuse and streaming support in bench. This allows us to optimize our infrastructure to provide top tier performance.

Infrastructure & Dependencies - Building a Scalable and Secure Technological Foundation - Future-Proofing our Investment

πŸ–₯ Infrastructure Updates - Ensuring Uninterrupted Service and Scalability for Accelerated AI Operations - A Cornerstone of Our Growth Strategy
  • Docker Base Images: Updated to the newest images, ensuring that our infrastructure remains compliant with industry standards.
  • Library Upgrades: Updating to the latest libraries. This is key to our growth and performance.

API & Configuration Updates - Enhancing Developer Productivity and Time-to-Market – Accelerating Innovation

πŸ”§ API Changes - Fostering Agility and Increasing Operational Efficiency
  • Config Object Definition: Standardized configuration object definition, simplifying the deployment of LLMs.
  • Decoder Interface Refinement: Improved code modularity.
  • Torch Compile Configuration: Enhanced to maximize inference speed.
  • Redundancy Removal: Removed code redundancy.
  • API Access Refinement: Improved API Access.

Fixed Issues: Bug Fixes & Stability Improvements - Maintaining Unwavering Product Quality and User Trust - A Foundation for Sustained Growth

πŸ›  Resolved Issues - Demonstrating our Commitment to Quality, Reliability, and User Satisfaction - Delivering Excellence
  • Fixed disaggregated service hang with MNNVL two-shot AllReduce (#4678).
  • Fixed EP load balancer misrouting (#4767).
  • Fixed cuda graph padding for spec decoding (#4853).
  • Fixed LLaMA 4 long context issue (#4809).
  • Fixed max_num_sequences calculation with overlap scheduling (#4532).
  • Fixed chunked prefill + overlap scheduling (#5761).
  • Fixed trtllm-bench hang due to LLM API IPC (#4798).
  • Fixed index out of bounds in spec decoding (#5954).
  • Fixed MTP illegal memory access in cuda graph warmup (#5947).
  • Fixed slot allocation error in spec decode + disagg (#5975).
  • Fixed attention window one-off bug for Gemma3 1B (#5564).
  • TikTok connection issue fixed
  • App performance issue fixed
  • Multi-model selection not visible on mobile for all themes fixed
  • AI Chat Pro UI issue fixed
  • GPT-5 responses show unnecessary numbers fixed
  • GPT responds with the same response twice fixed
  • AI Chat Pro Image Generation shows unknown numbers fixed
  • Voice Bot conversations not scoped to individual users fixed
  • AI Chat displays 2 models instead of one fixed
  • External Chatbot message alignment issue fixed

Known Issues & Limitations - Maintaining Full Transparency and Proactive Risk Management – Demonstrating Sound Governance

⚠️ Known Issues - Proactive Mitigation Strategies in Place – Mitigating Risk, Ensuring Long-Term Viability
  • accuracy/test_cli_flow::TestGpt2::test_beam_search_large is broken. Investigating and will resolve in 9.4.2.
  • Disaggregated serving + MTP + overlap scheduler together can cause accuracy issues. Actively working on a solution, which we expect to release in 9.4.2.
  • LLaMA4 functional support (>8K seq length) introduces a performance regression on Hopper. Addressing the performance regression on Hopper architecture will be prioritized, and optimized kernels will be created for the next release.