@blog.google
//
Google is expanding access to its Gemini AI app to all Google Workspace for Education users, marking a significant step in integrating AI into educational settings. This rollout, announced on June 20, 2025, provides educators and students with a range of AI-powered tools. These tools include real-time support for learning, assistance in creating lesson plans, and capabilities for providing feedback on student work, all designed to enhance the learning experience and promote AI literacy. The Gemini app is covered under the Google Workspace for Education Terms of Service, ensuring enterprise-grade data protection and compliance with regulations like FERPA, COPPA, FedRamp, and HIPAA.
A key aspect of this expansion is the implementation of stricter content policies for users under 18. These policies are designed to prevent potentially inappropriate or harmful responses, creating a safer online environment for younger learners. Additionally, Google is introducing a youth onboarding experience with AI literacy resources, endorsed by ConnectSafely and the Family Online Safety Institute, to guide students in using AI responsibly. The first time a user asks a fact-based question, a "double-check response" feature, powered by Google Search, will automatically run to validate the answer. Gemini incorporates LearnLM, Google’s family of models fine-tuned for learning and built with experts in education, making it a leading model for educational purposes. To ensure responsible use, Google provides resources for educators, including a Google teacher center offering guidance on incorporating Gemini into lesson plans and teaching responsible AI practices. Administrators can manage user access to the Gemini app through the Google Workspace Admin Help Center, allowing them to set up groups or organizational units to control access within their domain and tailor the AI experience to specific educational needs. References :
Classification:
Steve Vandenberg@Microsoft Security Blog
//
Microsoft is making significant strides in AI and data security, demonstrated by recent advancements and reports. The company's commitment to responsible AI is highlighted in its 2025 Responsible AI Transparency Report, detailing efforts to build trustworthy AI technologies. Microsoft is also addressing the critical issue of data breach reporting, offering solutions like Microsoft Data Security Investigations to assist organizations in meeting stringent regulatory requirements such as GDPR and SEC rules. These initiatives underscore Microsoft's dedication to ethical and secure AI development and deployment across various sectors.
AI's transformative potential is being explored in higher education, with Microsoft providing AI solutions for creating AI-ready campuses. Institutions are focusing on using AI for unique differentiation and innovation rather than just automation and cost savings. Strategies include establishing guidelines for responsible AI use, fostering collaborative communities for knowledge sharing, and partnering with technology vendors like Microsoft, OpenAI, and NVIDIA. Comprehensive training programs are also essential to ensure stakeholders are proficient with AI tools, promoting a culture of experimentation and ethical AI practices. Furthermore, Microsoft Research has achieved a breakthrough in computational chemistry by using deep learning to enhance the accuracy of density functional theory (DFT). This advancement allows for more reliable predictions of molecular and material properties, accelerating scientific discovery in fields such as drug development, battery technology, and green fertilizers. By generating vast amounts of accurate data and using scalable deep-learning approaches, the team has overcome limitations in DFT, enabling the design of molecules and materials through computational simulations rather than relying solely on laboratory experiments. References :
Classification:
@pub.towardsai.net
//
DeepSeek's R1 model is garnering attention as a potential game-changer for entrepreneurs, offering advancements in "reasoning per dollar." This refers to the amount of reasoning power one can obtain for each dollar spent, potentially unlocking opportunities previously deemed too expensive or technologically challenging. The model's high-reasoning capabilities at a reasonable cost are seen as a way to make advanced AI more accessible, particularly for tasks that require deep understanding and synthesis of information. One example is the creation of sophisticated AI-powered tools, like a "lawyer agent" that can review contracts, which were once cost-prohibitive.
The DeepSeek R1 model has been updated and released on Hugging Face, reportedly featuring significant changes and improvements. The update comes amidst both excitement and apprehension regarding the model's capabilities. While the model demonstrates promise in areas like content generation and customer support, concerns exist regarding potential political bias and censorship. This stems from observations of alleged Chinese government influence in the model's system instructions, which may impact the neutrality of generated content. The adoption of DeepSeek R1 requires careful self-assessment by businesses and individuals, weighing its strengths and potential drawbacks against specific needs and values. Users must consider the model's alignment with their data governance, privacy requirements, and ethical principles. For instance, while the model's content generation capabilities are strong, some categories might be censored or skewed by built-in constraints. Similarly, its chatbot integration may lead to heavily filtered replies, raising concerns about alignment with corporate values. Therefore, it is essential to be comfortable with the possible official or heavily filtered replies, and to consider monitoring the AI's responses to ensure they align with the business' values. References :
Classification:
|
BenchmarksBlogsResearch Tools |