A weekly Newsletter on technology applications in investment management with an AI / LLM and automation angle. Curated news, announcements, and posts, primarily directly from sources. We apply some of the ROI's Kubro(TM) Engine tools at the backend for production, yet with a human in the loop (for now). It's a start, and it will evolve on a weekly basis.
Disclaimers: content not fully human-verified, with AI summaries below. AI/LLMs may hallucinate and provide inaccurate summaries. Select items only, not intended as a comprehensive view. For information purposes only. Please DM with feedback and requests
The article presents a methodology for extracting structured risk factors from corporate 10-K filings using a three-stage pipeline: LLM extraction with supporting quotes, embedding-based semantic mapping to taxonomy categories, and LLM-as-a-judge validation to filter spurious assignments. The approach was evaluated by extracting 10,688 risk factors from S&P 500 companies and analyzing risk profile similarity across industry clusters. Autonomous taxonomy maintenance is introduced, with an AI agent achieving a 104.7% improvement in embedding separation in a case study. External validation shows same-industry companies have 63% higher risk profile similarity than cross-industry pairs (Cohen's d=1.06, AUC 0.82, p<0.001).
🔗 Source: arxiv.org View Source | Found on Jan 22, 2026
The study by Zefeng Chen and Darcy Pu investigates whether fully agentic AI can nowcast stock returns by autonomously evaluating Russell 1000 stocks daily, beginning in April 2025 when real-time web search became possible. Their framework is out-of-sample, free of look-ahead bias, temporally irreproducible, and entirely agentic—AI independently searches and synthesizes information without curated input. Results show that longing the top 20 ranked stocks yields a daily Fama-French five-factor plus momentum alpha of 18.4 basis points and an annualized Sharpe ratio of 2.43, with transaction costs under 10% of gross alpha; predictability is concentrated among top winners only.
🔗 Source: arxiv.org View Source | Found on Jan 21, 2026
Jan Rosenzweig's paper, "Fast Times, Slow Times: Timescale Separation in Financial Timeseries Data," introduces a method for separating multiscale processes in financial time series using variance and tail stationarity criteria, formulated as generalized eigenvalue problems. The technique enables identification of slow and fast components in asset returns and prices, with applications to parameter drift, mean reversion, and tail risk management. Empirical examples are provided using currencies, equity ETFs, and treasury yields to demonstrate the practical utility of the approach. The paper was submitted on 16 January 2026 to arXiv under portfolio management subjects.
🔗 Source: arxiv.org View Source | Found on Jan 19, 2026
The article, "Beyond Visual Realism: Toward Reliable Financial Time Series Generation" by Fan Zhang and colleagues, addresses the limitations of generative models such as GANs and WGAN-GP in producing usable synthetic financial time series data. The authors identify that these models often fail under trading backtests due to neglecting financial asymmetry and rare tail events. To solve this, they propose the Stylized Facts Alignment GAN (SFAG), which incorporates differentiable structural constraints based on key stylized facts. Experiments on Shanghai Composite Index data from 2004 to 2024 demonstrate that SFAG generates more robust and realistic synthetic series for momentum strategy evaluation.
🔗 Source: arxiv.org View Source | Found on Jan 21, 2026
The paper introduces the Determinism-Faithfulness Assurance Harness (DFAH), a framework for measuring trajectory determinism and evidence-conditioned faithfulness in tool-using LLM agents deployed in financial services. In non-agentic baseline experiments across 74 configurations involving 12 models and 4 providers, 7-20B parameter models achieved 100% determinism, while models with over 120B parameters required validation samples that were 3.7 times larger for equivalent reliability. A positive Pearson correlation (r = 0.45, p < 0.01, n = 51 at T=0.0) was found between determinism and faithfulness. Three financial benchmarks and an open-source stress-test harness are provided.
🔗 Source: arxiv.org View Source | Found on Jan 23, 2026
The article introduces Look-Ahead-Bench, a standardized benchmark designed to measure look-ahead bias in Point-in-Time (PiT) Large Language Models (LLMs) for finance. The benchmark evaluates model behavior in practical financial scenarios and analyzes performance decay across different market regimes using quantitative baselines. Prominent open-source LLMs, including Llama 3.1 (8B and 70B) and DeepSeek 3.2, are compared with PiT-Inference models (Pitinf-Small, Pitinf-Medium, Pitinf-Large). Results show significant lookahead bias in standard LLMs as measured by alpha decay, while Pitinf models demonstrate improved generalization and reasoning with increased size. Code is available on GitHub.
🔗 Source: arxiv.org View Source | Found on Jan 21, 2026
According to the sixth annual “NVIDIA State of AI in Financial Services” report, based on a survey of over 800 industry professionals, 89% said AI is helping increase annual revenue and decrease costs, with 64% reporting revenue increases over 5% and 29% over 10%, while 61% saw cost reductions above 5%. Nearly all respondents expect their AI budgets to increase or remain steady in the coming year. Active AI use rose to 65%, up from 45% last year, with generative AI usage at 61%. Open source models are important for 84%, and agentic AI has been deployed by 21%, with another 22% planning deployment.
🔗 Source: blogs.nvidia.com View Source | Found on Jan 22, 2026
Qwen3-TTS is a speech generation system developed by Qwen, featuring voice clone, voice design, high-quality human-like speech generation, and natural language-based voice control. It uses the Qwen3-TTS-Tokenizer-12Hz multi-codebook speech encoder for efficient compression and robust representation of speech signals, preserving paralinguistic and acoustic features. The lightweight non-DiT architecture enables high-speed, high-fidelity reconstruction with Dual-Track modeling for rapid bidirectional streaming. The open-sourced model series includes 1.7B and 0.6B sizes, supports 10 languages plus dialects, adapts tone and emotion contextually, improves robustness to text noise, and is available on GitHub and via the Qwen API.
🔗 Source: qwen.ai View Source | Found on Jan 22, 2026
The article announces the addition of a feature enabling Anthropic-compatible applications, such as Claude Code, to work with local models via llama-server. Contributed by noname22 in PR #17570, the implementation converts Anthropic's format to OpenAI internally and utilizes llama.cpp's existing inference pipeline for performance benefits with quantized models. Users can point their Anthropic client to the endpoint and start the server with any GGUF compatible models for tool use support. For agentic workloads, specialized coding models like Nemotron, Qwen3 Coder, Kimi K2, or MiniMax M2 are recommended.
🔗 Source: huggingface.co View Source | Found on Jan 19, 2026
Anthropic has published a new constitution for its AI model, Claude, detailing the values and behaviors it aims to instill. Released under a Creative Commons CC0 1.0 Deed, the constitution is central to Claude’s training and prioritizes broad safety, ethics, compliance with Anthropic’s guidelines, and genuine helpfulness in that order. The document provides guidance on handling complex situations and tradeoffs, outlines hard constraints such as prohibiting assistance with bioweapons attacks, and emphasizes transparency by making the constitution publicly available. Feedback from external experts informed its development, and updates will be maintained on Anthropic’s website.
🔗 Source: anthropic.com View Source | Found on Jan 21, 2026
On January 22, Baidu released the officially launched native all-modal large model Wenxin 5.0, featuring 24 trillion parameters and unified modeling technology for text, image, audio, and video inputs/outputs. Wenxin 5.0’s language and multimodal understanding capabilities surpassed Gemini-2.5-Pro and GPT-5-High in over 40 benchmark evaluations, ranking among the global leaders. The Qianfan platform enables enterprise users and developers to easily access Wenxin 5.0; over 1.3 million Agents have been developed on Qianfan with daily tool calls exceeding ten million times. In 2025, Baidu unveiled China’s first self-developed Kunlun chip cluster of thirty thousand cards and new M100/M300 chips.
🔗 Source: cloud.baidu.com View Source | Found on Jan 23, 2026
In December 2025, President Trump signed an Executive Order blocking state-level AI laws deemed incompatible with a minimally burdensome national policy framework, and a bill has been introduced to counter this order. The European Commission published a Digital Omnibus on AI Regulation Proposal in November 2025 to simplify the EU AI Act and delay high-risk system application dates, pending European Parliament approval by August 2026. Korea’s Basic AI Act and Vietnam’s first dedicated AI law will take effect in 2026. Multiple countries have enacted risk-based AI laws targeting sectors like employment, education, and essential services, with generative AI regulations addressing deepfakes in India, the UK, and Denmark.
🔗 Source: holisticai.com View Source | Found on Jan 19, 2026
At the 2026 World Economic Forum in Davos, NVIDIA CEO Jensen Huang described AI as the foundation of “the largest infrastructure buildout in human history,” encompassing energy, chips, data centers, AI models, and applications. Huang stated that 2025 saw over $100 billion in global venture capital investment—most into AI-native startups—driving job creation across sectors such as healthcare, manufacturing, and construction. He emphasized that AI increases demand for skilled labor and improves productivity in fields like radiology and nursing. Huang also highlighted AI’s accessibility, its potential to close technology divides, and urged nations to treat AI as essential infrastructure.
🔗 Source: blogs.nvidia.com View Source | Found on Jan 21, 2026
The Sanctioned Securities Data File (SSDF) from LSEG Risk Intelligence is an instrument-level dataset that links global sanctions designations and ownership and control relationships directly to financial securities, enabling financial institutions to identify and manage sanctions exposure. Approximately one-third of sanctions-linked securities are connected through ownership and control pathways, while around 60% of these instruments remain actively traded. Rights and entitlements, debt instruments, and structured products account for about 80% of sanctions-linked issuance. Russia-imposed measures represent the largest share of identified exposure, with significant contributions from the EU, US, New Zealand, and Ukraine.
🔗 Source: lseg.com View Source | Found on Jan 20, 2026
AI capital expenditures (capex) are projected to boost US GDP growth by approximately 140 basis points in 2026 and 150 basis points in 2027, primarily due to large-scale data center build-outs, with significant profits accruing to companies like Nvidia. However, this investment is not expected to meaningfully support labor markets, as data centers require relatively few workers compared to other sectors; for example, Meta’s $1.5 billion El Paso data center will support only 100 jobs versus AESC’s $1.6 billion battery plant supporting 1,620 jobs. The AI capex boom is also creating acute price pressures on power, memory chips, and materials such as copper and aluminum.
🔗 Source: bridgewater.com View Source | Found on Jan 21, 2026
In 2026, Two Sigma is integrating advanced AI models and tools across all workflows, including alpha identification and legal compliance, to enhance productivity and outcomes. The firm emphasizes making AI systems “Two Sigma aware” for better understanding of internal processes. The focus in AI development is shifting from scaling size to improving efficiency through new architectures and neural compression. Interpretability of models is becoming a competitive priority, while multimodal models are enabling unified representations for trading environments. Despite productivity gains—such as rapidly generating predictive features—challenges like overfitting and regime changes persist, requiring rigorous research discipline and skilled talent.
🔗 Source: twosigma.com View Source | Found on Jan 21, 2026
Global equities reached new highs in 2025 despite trade and tariff challenges, with the “Magnificent 7” tech giants leading gains but broader earnings also surprising positively. AI infrastructure development drove growth, especially in datacenter equipment, which is expected to see around 25% annual growth for four to five years due to supply constraints. Semiconductor capital equipment spending is projected to rise in 2026, supported by memory and advanced logic investments. Consumer discretionary spending remains robust above pre-pandemic levels, though lower-income consumers face greater financial strain. Healthcare saw recovery after pricing agreements and a 15% cap on pharmaceutical imports were established.
🔗 Source: pinebridge.com View Source | Found on Jan 19, 2026