The post GitHub Copilot Enhances Code Search with New Embedding Model appeared on BitcoinEthereumNews.com. Ted Hisokawa Sep 26, 2025 03:41 GitHub introduces a new Copilot embedding model, enhancing code search in VS Code with improved accuracy and efficiency, according to GitHub’s announcement. GitHub has announced a significant upgrade to its Copilot tool, introducing a new embedding model that promises to enhance code search within Visual Studio Code (VS Code). This development aims to make code retrieval faster, more memory-efficient, and significantly more accurate, as detailed in a recent GitHub blog post. Enhanced Code Retrieval The new Copilot embedding model brings a 37.6% improvement in retrieval quality, doubling the throughput and reducing the index size by eight times. This means developers can expect more accurate code suggestions, faster response times, and reduced memory usage in VS Code. The model effectively provides the correct code snippets needed, minimizing irrelevant results. Why the Upgrade Matters Efficient code search is crucial for a seamless AI coding experience. Embeddings, which are vector representations, play a key role in retrieving semantically relevant code and natural language content. The improved embeddings result in higher retrieval quality, thereby enhancing the overall GitHub Copilot experience. Technical Improvements GitHub has trained and deployed this new model specifically for code and documentation, enhancing context retrieval for various Copilot modes. The update has shown significant improvements, with C# developers experiencing a 110.7% increase in code acceptance ratios and Java developers seeing a 113.1% rise. Training and Evaluation The model was optimized using contrastive learning techniques, such as InfoNCE loss and Matryoshka Representation Learning, to improve retrieval quality. A key aspect of the training involved using ‘hard negatives’—code examples that appear correct but are not—helping the model distinguish between nearly correct and actually correct code snippets. Future Prospects GitHub plans to expand its training and evaluation data to include… The post GitHub Copilot Enhances Code Search with New Embedding Model appeared on BitcoinEthereumNews.com. Ted Hisokawa Sep 26, 2025 03:41 GitHub introduces a new Copilot embedding model, enhancing code search in VS Code with improved accuracy and efficiency, according to GitHub’s announcement. GitHub has announced a significant upgrade to its Copilot tool, introducing a new embedding model that promises to enhance code search within Visual Studio Code (VS Code). This development aims to make code retrieval faster, more memory-efficient, and significantly more accurate, as detailed in a recent GitHub blog post. Enhanced Code Retrieval The new Copilot embedding model brings a 37.6% improvement in retrieval quality, doubling the throughput and reducing the index size by eight times. This means developers can expect more accurate code suggestions, faster response times, and reduced memory usage in VS Code. The model effectively provides the correct code snippets needed, minimizing irrelevant results. Why the Upgrade Matters Efficient code search is crucial for a seamless AI coding experience. Embeddings, which are vector representations, play a key role in retrieving semantically relevant code and natural language content. The improved embeddings result in higher retrieval quality, thereby enhancing the overall GitHub Copilot experience. Technical Improvements GitHub has trained and deployed this new model specifically for code and documentation, enhancing context retrieval for various Copilot modes. The update has shown significant improvements, with C# developers experiencing a 110.7% increase in code acceptance ratios and Java developers seeing a 113.1% rise. Training and Evaluation The model was optimized using contrastive learning techniques, such as InfoNCE loss and Matryoshka Representation Learning, to improve retrieval quality. A key aspect of the training involved using ‘hard negatives’—code examples that appear correct but are not—helping the model distinguish between nearly correct and actually correct code snippets. Future Prospects GitHub plans to expand its training and evaluation data to include…

GitHub Copilot Enhances Code Search with New Embedding Model

2025/09/27 19:29


Ted Hisokawa
Sep 26, 2025 03:41

GitHub introduces a new Copilot embedding model, enhancing code search in VS Code with improved accuracy and efficiency, according to GitHub’s announcement.





GitHub has announced a significant upgrade to its Copilot tool, introducing a new embedding model that promises to enhance code search within Visual Studio Code (VS Code). This development aims to make code retrieval faster, more memory-efficient, and significantly more accurate, as detailed in a recent GitHub blog post.

Enhanced Code Retrieval

The new Copilot embedding model brings a 37.6% improvement in retrieval quality, doubling the throughput and reducing the index size by eight times. This means developers can expect more accurate code suggestions, faster response times, and reduced memory usage in VS Code. The model effectively provides the correct code snippets needed, minimizing irrelevant results.

Why the Upgrade Matters

Efficient code search is crucial for a seamless AI coding experience. Embeddings, which are vector representations, play a key role in retrieving semantically relevant code and natural language content. The improved embeddings result in higher retrieval quality, thereby enhancing the overall GitHub Copilot experience.

Technical Improvements

GitHub has trained and deployed this new model specifically for code and documentation, enhancing context retrieval for various Copilot modes. The update has shown significant improvements, with C# developers experiencing a 110.7% increase in code acceptance ratios and Java developers seeing a 113.1% rise.

Training and Evaluation

The model was optimized using contrastive learning techniques, such as InfoNCE loss and Matryoshka Representation Learning, to improve retrieval quality. A key aspect of the training involved using ‘hard negatives’—code examples that appear correct but are not—helping the model distinguish between nearly correct and actually correct code snippets.

Future Prospects

GitHub plans to expand its training and evaluation data to include more languages and repositories. The company is also refining its hard negative mining pipeline to enhance quality further, with goals to deploy larger, more accurate models leveraging the efficiency gains from this update.

This latest enhancement is a step towards making AI coding assistants more reliable and efficient for developers, promising a smarter and more dependable tool for everyday development.

Image source: Shutterstock


Source: https://blockchain.news/news/github-copilot-enhances-code-search-with-new-embedding-model

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact [email protected] for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

The Future of Secure Messaging: Why Decentralization Matters

The Future of Secure Messaging: Why Decentralization Matters

The post The Future of Secure Messaging: Why Decentralization Matters appeared on BitcoinEthereumNews.com. From encrypted chats to decentralized messaging Encrypted messengers are having a second wave. Apps like WhatsApp, iMessage and Signal made end-to-end encryption (E2EE) a default expectation. But most still hinge on phone numbers, centralized servers and a lot of metadata, such as who you talk to, when, from which IP and on which device. That is what Vitalik Buterin is aiming at in his recent X post and donation. He argues the next steps for secure messaging are permissionless account creation with no phone numbers or Know Your Customer (KYC) and much stronger metadata privacy. In that context he highlighted Session and SimpleX and sent 128 Ether (ETH) to each to keep pushing in that direction. Session is a good case study because it tries to combine E2E encryption with decentralization. There is no central message server, traffic is routed through onion paths, and user IDs are keys instead of phone numbers. Did you know? Forty-three percent of people who use public WiFi report experiencing a data breach, with man-in-the-middle attacks and packet sniffing against unencrypted traffic among the most common causes. How Session stores your messages Session is built around public key identities. When you sign up, the app generates a keypair locally and derives a Session ID from it with no phone number or email required. Messages travel through a network of service nodes using onion routing so that no single node can see both the sender and the recipient. (You can see your message’s node path in the settings.) For asynchronous delivery when you are offline, messages are stored in small groups of nodes called “swarms.” Each Session ID is mapped to a specific swarm, and your messages are stored there encrypted until your client fetches them. Historically, messages had a default time-to-live of about two weeks…
Share
BitcoinEthereumNews2025/12/08 14:40
Grayscale Files Sui Trust as 21Shares Launches First SUI ETF Amid Rising Demand

Grayscale Files Sui Trust as 21Shares Launches First SUI ETF Amid Rising Demand

The post Grayscale Files Sui Trust as 21Shares Launches First SUI ETF Amid Rising Demand appeared on BitcoinEthereumNews.com. The Grayscale Sui Trust filing and 21Shares’ launch of the first SUI ETF highlight surging interest in regulated Sui investments. These products offer investors direct exposure to the SUI token through spot-style structures, simplifying access to the Sui blockchain’s growth without direct custody needs, amid expanding altcoin ETF options. Grayscale’s spot Sui Trust seeks to track SUI price performance for long-term holders. 21Shares’ SUI ETF provides leveraged exposure, targeting traders with 2x daily returns. Early trading data shows over 4,700 shares exchanged, with volumes exceeding $24 per unit in the debut session. Explore Grayscale Sui Trust filing and 21Shares SUI ETF launch: Key developments in regulated Sui investments for 2025. Stay informed on altcoin ETF trends. What is the Grayscale Sui Trust? The Grayscale Sui Trust is a proposed spot-style investment product filed via S-1 registration with the U.S. Securities and Exchange Commission, aimed at providing investors with direct exposure to the SUI token’s price movements. This trust mirrors the performance of SUI, the native cryptocurrency of the Sui blockchain, minus applicable fees, offering a regulated avenue for long-term participation in the network’s ecosystem. By holding SUI assets on behalf of investors, it eliminates the need for individuals to manage token storage or transactions directly. ⚡ LATEST: GRAYSCALE FILES S-1 FOR $SUI TRUSTThe “Grayscale Sui Trust,” is a spot-style ETF designed to provide direct exposure to the $SUI token. Grayscale’s goal is to mirror SUI’s market performance, minus fees, giving long-term investors a regulated, hassle-free way to… pic.twitter.com/mPQMINLrYC — CryptosRus (@CryptosR_Us) December 6, 2025 How does the 21Shares SUI ETF differ from traditional funds? The 21Shares SUI ETF, launched under the ticker TXXS, introduces a leveraged approach with 2x daily exposure to SUI’s price fluctuations, utilizing derivatives for amplified returns rather than direct spot holdings. This structure appeals to short-term…
Share
BitcoinEthereumNews2025/12/08 14:20