10th Indian Delegation to Dubai, Gitex & Expand North Star – World’s Largest Startup Investor Connect
Artificial Intelligence

GitHub’s Copilot Enterprise is now generally available at $39 a month


GitHub today announced the general availability of Copilot Enterprise, the $39/month version of its code completion tool and developer-centric chatbot for large businesses. Copilot Enterprise includes all of the features of the existing Business plan, including IP indemnity, but extends this with a number of crucial features for larger teams. The highlight here is the ability to reference an organization’s internal code and knowledge base. Copilot is now also integrated with Microsoft’s Bing search engine (currently in beta) and soon, users will also be able to fine-tune Copilot’s models based on a team’s existing codebase as well.

With that, new developers on a team can, for example, ask Copilot how to deploy a container image to the cloud and get an answer that is specific to the process in their organization. For a lot of developers, after all, it’s not necessarily understanding the codebase that is a roadblock to being productive when moving companies but understanding the different processes — though Copilot can obviously help with understanding the code, too.

Image Credits: GitHub

Many teams already keep their documentation in GitHub repositories today, making it relatively easy for Copilot to reason over it. Indeed, as GitHub CEO Thomas Dohmke told me, since GitHub itself stores virtually all of its internal documents on the service — and recently gave access to these new features to all of its employees — some people have started using it for non-engineering questions, too, and started asking Copilot about vacation policies, for example.

Dohmke told me that customers had been asking for these features to reference internal information from the earliest days of Copilot. “A lot of the things that developers do within organizations are different to what they do at home or in open source, in the sense that organizations have a process or a certain library to use — and many of them have internal tools, systems and dependencies that do not exist like that on the outside,” he noted.

As for the Bing integration, Dohmke noted that this would be useful for asking Copilot about things that may have changed since the model was originally trained (think open source libraries or APIs). For now, this feature is only available in the Enterprise version and while Dohmke wouldn’t say much about whether it will come to other editions as well, I wouldn’t be surprised if GitHub brought this capability to the other tiers at a later point, too.

Image Credits: GitHub

One feature that will likely remain an enterprise feature — in part because of its associated cost — is fine-tuning, which will launch soon. “We let companies pick a set of repositories in their GitHub organization and then fine-tune the model on those repositories,” Dohmke explained. “We’re abstracting the complexity of generative AI and fine-tuning away from the customer and let them leverage their codebase to generate an optimized model for them that then is used within the Copilot scenarios.” He did note that this also means that the model can’t be as up-to-date as when using embeddings, skills and agents (like the new Bing agent). He argues that all of this is complementary, though, and the customers who are already testing this feature are seeing significant improvements. That’s especially true for teams that are working with codebases in languages that aren’t as widely used as the likes of Python and JavaScript, or with internal libraries that don’t really exist outside of an organization.

On top of talking about today’s release, I also asked Dohmke about his high-level thinking of where Copilot is going next. The answer is essentially “more Copilot in more places. I think, in the next year, we’re going to see an increasing focus on that end-to-end experience of putting Copilots where you already do the work as opposed to creating a new destination to go and copy and paste stuff there. I think that’s where we at GitHub are incredibly excited about the opportunity that we have by putting Copilot on github.com by having Copilot available in the place where developers are already collaborating, where they’re already building the world’s software.”

Image Credits: GitHub

Talking about the underlying technology and where that is going, Dohmke noted that the auto-completion feature currently runs on GPT 3.5 Turbo. Because of its latency requirements, GitHub never moved that model to GPT 4, but Dohmke also noted the team has updated the model “more than half a dozens times” since the launch of Copilot Business.

As of now, it doesn’t look like GitHub will follow the Google model of differentiating its pricing tiers by the size of the models that power those experiences. “Different use cases require different models. Different optimizations — latency, accuracy, quality of the outcome, responsible AI — for each model version play a big role to make sure that the output is ethical, compliant and secure and doesn’t generate a lower-quality code than what our customers expect. We will continue going down that path of using the best models for the different pieces of the Copilot experience,” Dohmke said.



Source link

by Team SNFYI

Facebook is testing a new feature that invites some users—mainly in the US and Canada—to let Meta AI access parts of their phone’s camera roll. This opt-in “cloud processing” option uploads recent photos and videos to Meta’s servers so the AI can offer personalized suggestions, such as creating collages, highlight reels, or themed memories like birthdays and graduations. It can also generate AI-based edits or restyles of those images. Meta says this is optional and assures users that the uploaded media won’t be used for advertising. However, to enable this, people must agree to let Meta analyze faces, objects, and metadata like time and location. Currently, the company claims these photos won’t be used to train its AI models—but they haven’t completely ruled that out for the future. Typically, only the last 30 days of photos get uploaded, though special or older images might stay on Meta’s servers longer for specific features. Users have the option to disable the feature anytime, which prompts Meta to delete the stored media after 30 days. Privacy experts are concerned that this expands Meta’s reach into private, unpublished images and could eventually feed future AI training. Unlike Google Photos, which explicitly states that user photos won’t train its AI, Meta hasn’t made that commitment yet. For now, this is still a test run for a limited group of people, but it highlights the tension between AI-powered personalization and the need to protect personal data.

by Team SNFYI

News Update Bymridul     |    March 14, 2024 Meesho, an online shopping platform based in Bengaluru, has announced its largest Employee Stock Ownership Plan (ESOP) buyback pool to date, totaling Rs 200 crore. This buyback initiative extends to both current and former employees, providing wealth creation opportunities for approximately 1,700 individuals. Ashish Kumar Singh, Meesho’s Chief Human Resources Officer, emphasized the company’s commitment to rewarding its teams, stating, “At Meesho, our employees are the driving force behind our success.” Singh further highlighted the company’s dedication to providing opportunities for wealth creation despite prevailing macroeconomic conditions. This marks the fourth wealth generation opportunity at Meesho, with the size of the buyback program increasing each year. In previous years, Meesho conducted buybacks worth over Rs 8.2 crore in February 2020, Rs 41.4 crore in November 2020, and Rs 45.5 crore in October 2021. Meesho’s profitability journey began in July 2023, making it the first horizontal Indian e-commerce company to achieve profitability. Despite turning profitable, Meesho continues to maintain positive cash flow and focuses on enhancing efficiencies across various cost items. The company’s revenue from operations for FY 2022-23 witnessed a remarkable growth of 77% over the previous year, amounting to Rs 5,735 crore. This growth can be attributed to Meesho’s leadership position as the most downloaded shopping app in India in both 2022 and 2023, increased transaction frequency among existing customers, and a diversified category mix. Additionally, Meesho’s focus on improving monetization through value-added seller services contributed to its revenue growth. Meesho also disclosed its audited performance for the first half of FY 2023-24, reporting consolidated revenues from operations of Rs 3,521 crore, marking a 37% year-over-year increase. The company achieved profitability in Q2 FY24, with a significant reduction in losses compared to the previous year. Furthermore, Meesho recorded impressive app download numbers, reaching 145 million downloads in India in 2023 and surpassing 500 million downloads in H1 FY 2023-24. Follow Startup Story Source link

by Team SNFYI

You might’ve heard of Grok, X’s answer to OpenAI’s ChatGPT. It’s a chatbot, and, in that sense, behaves as as you’d expect — answering questions about current events, pop culture and so on. But unlike other chatbots, Grok has “a bit of wit,” as X owner Elon Musk puts it, and “a rebellious streak.” Long story short, Grok is willing to speak to topics that are usually off limits to other chatbots, like polarizing political theories and conspiracies. And it’ll use less-than-polite language while doing so — for example, responding to the question “When is it appropriate to listen to Christmas music?” with “Whenever the hell you want.” But Grok’s ostensible biggest selling point is its ability to access real-time X data — an ability no other chatbots have, thanks to X’s decision to gatekeep that data. Ask it “What’s happening in AI today?” and Grok will piece together a response from very recent headlines, while ChatGPT, by contrast, will provide only vague answers that reflect the limits of its training data (and filters on its web access). Earlier this week, Musk pledged that he would open source Grok, without revealing precisely what that meant. So, you’re probably wondering: How does Grok work? What can it do? And how can I access it? You’ve come to the right place. We’ve put together this handy guide to help explain all things Grok. We’ll keep it up to date as Grok changes and evolves. How does Grok work? Grok is the invention of xAI, Elon Musk’s AI startup — a startup reportedly in the process of raising billions in venture capital. (Developing AI’s expensive.) Underpinning Grok is a generative AI model called Grok-1, developed over the course of months on a cluster of “tens of thousands” of GPUs (according to an xAI blog post). To train it, xAI sourced data both from the web (dated up to Q3 2023) and feedback from human assistants that xAI refers to as “AI tutors.” On popular benchmarks, Grok-1 is about as capable as Meta’s open source Llama 2 chatbot model and surpasses OpenAI’s GPT-3.5, xAI claims. Image Credits: xAI Human-guided feedback, or reinforcement learning from human feedback (RLHF), is the way most AI-powered chatbots are fine-tuned these days. RLHF involves training a generative model, then gathering additional information to train a “reward” model and fine-tuning the generative model with the reward model via reinforcement learning. RLHF is quite good at “teaching” models to follow instructions — but not perfect. Like other models, Grok is prone to hallucinating, sometimes offering misinformation and false timelines when asked about news. And these can be severe — like wrongly claiming that the Israel–Palestine conflict reached a ceasefire when it hadn’t. For questions that stretch beyond its knowledge base, Grok leverages “real-time access” to info on X (and from Tesla, according to Bloomberg). And, similar to ChatGPT, the model has internet browsing capabilities, enabling it to search the web for up-to-date information about topics. Musk has promised improvements with the …