Tech news
Using space science to protect Saudi Arabia’s environment
Learning space science has delivered significant environmental benefits worldwide, helping many countries better understand and manage climate challenges.
Saudi Arabia is now taking steps not only to explore the galaxy but also to invest in future generations who can apply space science to pressing environmental issues at home.
Last November, the Space Academy, part of the Saudi Space Agency, launched a series of seminars designed to enhance knowledge and develop skills in space science and technology, with a particular focus on Earth observation.
Running for nearly a month, the program formed part of a broader strategy to nurture national talent, raise scientific awareness, and build data capabilities that support innovation and research across the Kingdom.
Developing space sector can eventually help reduce some of the critical climate issues such as drought and air pollution. (AFP)
As efforts to strengthen the sector continue, important questions remain: How can space science translate into tangible environmental benefits? And how large is the global space economy?
In an interview with Arab News, Fahad Alhussain, co-founder of SeedFord, highlighted the scale of the opportunity and its environmental impact.
“To be frank, the slogan that we always use in space is that ‘saving the Earth from the space.’ It is all about this,” Alhusain told Arab News.
You can recall a lot of related environmental issues like global warming, related to forests, related to the damage that happens to the environment. Without space, it would be almost impossible to see the magnitude of these damages.”
According to Alhussain, satellites have transformed how experts observe environmental changes on Earth, offering a comprehensive view that was previously impossible.
“By collecting data and using satellites… You can better analyze and measure so many things that help the environment,” said Fahad Alhussain. (Supplied)
He said that “the transformation of technology allows even the non-optical ways of measuring, assessing, and discovering what is going on in the environment … you can even anticipate fire before it happens in the forest.”
“You can detect the ice-melt down, you can get huge amount of information and can see it through the weather maps…there is a huge section in the economy for the environment,” Alhussain commented.
A 2022 report by Ryan Brukardt, a senior partner at McKinsey & Company, published by McKinsey Quarterly, found that more than 160 satellites currently monitor Earth to assess the impacts of global warming and detect activities such as illegal logging.
Brukardt cited NASA as an example of how advanced satellite tools are used to track environmental changes, including shifts in ocean conditions, cloud cover, and precipitation patterns. He also noted that satellite data can help governments determine when immediate action is needed, particularly in response to wildfires.
Story by Arab News
Tech news
xAI sues Colorado over AI law
Elon Musk’s artificial intelligence company, xAI, has filed a lawsuit against the state of Colorado over a new AI law set to take effect in June.
The suit seeks to block the state from enforcing the law, which would impose new requirements on AI systems to protect state residents from “algorithmic discrimination” in sectors such as education, employment, healthcare, housing and financial services.
Colorado was the first state to pass a comprehensive bill to regulate AI.
The company claims the law infringes on its first amendment free-speech protections and would force xAI to “promote the state’s ideological views on various matters, racial justice in particular”, according to the Financial Times, which first reported the lawsuit. “Its provisions prohibit developers of AI systems from producing speech that the state of Colorado dislikes.”
The lawsuit, filed in US district court in Colorado, comes as battles have raged at the state and federal level over how to regulate the fast-growing technology. States such as California and New York have been working to rein in AI with regulations, while the Trump administration has been trying to loosen the rules and place a moratorium on state laws.
xAI, which makes the chatbot Grok, has been plagued with accusations of discrimination. The chatbot has consistently spewed racist, sexist and antisemitic content, put forth conspiracies of “white genocide” and referred to itself as “MechaHitler”.
Katie Miller, a former spokesperson for xAI and the wife of Trump adviser Stephen Miller, heralded the lawsuit in a post on X on Thursday: “Colorado wants to force Grok to follow its views on equity and race, instead of being maximally truth-seeking. Grok answers to evidence, not woke leftist government regulations.”
Jared Polis, Colorado’s Democratic governor, signed the bill into law in 2024 but said it was “with reservations”. He has called on state legislators to amend it. The legislation was intended to go into effect in February, but was pushed until 30 June.
xAI, which merged with Musk’s rocket business SpaceX earlier this year, is seeking an injunction to block the enforcement of the Colorado law and a court declaration saying the legislation is unconstitutional.
The Colorado attorney general’s office declined to comment on the lawsuit and xAI did not return a request for comment.
The Guardian
Tech news
Abu Dhabi Unveils Low-Cost AI Model to Rival OpenAI, DeepSeek
A new challenger in the global artificial intelligence race has entered the ring.
The Mohamed bin Zayed University of Artificial Intelligence (MBZUAI), an AI-focused research university established by the United Arab Emirates, announced on Tuesday the release of a new, low-cost reasoning model to rival OpenAI and DeepSeek.
It comes after DeepSeek, a Chinese AI lab, earlier this year shocked the world with the release of a reasoning model called R1 which it said could outperform OpenAI but with far less training costs.
At just 32 billion parameters, MBZUAI’s model, dubbed K2 Think, is much smaller than competing systems from OpenAI and DeepSeek. It was built on top of Alibaba’s open-source Qwen 2.5 model and is run and tested on hardware provided by AI chipmaker Cerebas.
For context, DeepSeek’s R1 has a total of 671 billion parameters, which is essentially another term for the variables that an AI language model learns to understand and generate language. OpenAI doesn’t disclose the parameter counts of its AI models.
K2 Think was developed in partnership with G42, the buzzy UAE-based AI firm backed by U.S. tech giant Microsoft. The researchers behind it say it delivers performance on par with the flagship reasoning models of OpenAI and DeepSeek — despite being a fraction of the size.
They cited the benchmarks AIME24, AIME25, HMMT25 and OMNI-Math-HARD, which relate to math, coding benchmark LiveCodeBenchv5 and science benchmark GPQA-Diamond.
How did they do it?
Hector Liu, director of MBZUAI’s Institute of Foundation Models, told CNBC the team behind K2 Think were able to achieve such high levels of performance by using a number of methods.
They include long chain-of-thought (CoT) supervised fine-tuning — a method of step-by-step reasoning — as well as so-called test-time scaling, which is a technique for improving performance by allocating extra computing resources during “inferencing” — or, applying learned knowledge to data it’s never seen before.
“What was special about our model is we treat it more like a system than just a model,” Liu told CNBC. “So, unlike a regular open-source model where we can just release the model, we actually deploy the model and see how we can improve the model over time.”
“If you ask me which one of the single steps is the most important, it’s very hard to say. It’s more like a system method work where all these methods combined delivered the final result,” he added.
Why does it matter?
There are two countries on the world stage that stand out as the forerunners in the AI race: the U.S. and China.
America’s tech giants and startups like OpenAI led the early momentum with so-called foundation models, which aim to fulfill a wide range of tasks by relying on vast amounts of training data. However, DeepSeek’s breakthrough with R1 earlier this year reinforced China’s position as a formidable AI player in its own right.
More recently, the UAE has sought to position itself as a global leader in AI in a bid to enhance its geopolitical influence and diversify its economy beyond crude oil dependency.
The region can point to its AI development firm G42 as an example of how it’s gaining ground in the space. However, it faces fierce competition from neighboring Saudi Arabia, which is looking to develop full-stack AI capabilities via Humain, a company launched under the Public Investment Fund in May.
Beyond that, there are also geopolitical complexities that shroud the UAE’s AI ambitions. Microsoft’s investment and partnership with G42 last year attracted a great deal of scrutiny in the U.S. related to the company’s relationship with China.
More broadly, the UAE’s AI industry still has a long way to go to reach the scale of its U.S. and Chinese counterparts. OpenAI and the Big Tech players have enjoyed a good head start with their respective foundation AI models, while Beijing has long considered AI a strategic priority.
Focus on scientific breakthroughs
While K2 Think demonstrates performance on par with OpenAI, the system’s developers say the aim is not to build a chatbot like ChatGPT. Richard Morton, managing director for MBZUAI’s Institute of Foundation Models, explains the model is intended to serve specific uses in fields like math and science.
“The fact is that the fundamental reasoning of the human brain is the cornerstone of all the thinking process,” Morton told CNBC.
“With this particular application, instead of taking 1,000, 2,000 human beings five years to think through a particular question, or go through a particular set of clinical trials or something like that, this vastly condenses that period.”
It could also expand the reach of advanced AI technologies in regions that don’t have access to the kind of capital and infrastructure U.S. firms possess.
“What we’re discovering is that you can do a lot more with less,” Morton said.
CNBC
Tech news
Viral ChatGPT AI caricature trend
A new AI trend is flooding Instagram, LinkedIn, X, and WhatsApp. Users upload selfies and ask ChatGPT to turn them into cartoon-style versions of themselves at work — surrounded by laptops, notebooks, coffee mugs, or office backdrops.
The viral prompt is simple: “Create a caricature of me and my job based on everything you know about me.”
What’s new about these AI caricatures?
Unlike older portrait apps that just stylised photos, this trend adds contextual personalisation:
- Analyses patterns from past interactions — tone, frequently discussed topics, interests, hobbies.
- Generates a cartoon scene reflecting your professional and personal life.
- Includes tools of your trade, subtle lifestyle cues, and workplace details.
Why it works:
These images go beyond simple cartoons — they reflect identity, profession, and mood, making them ideal for social media profiles.
The trend works because anyone can join in, no design skills or editing apps needed, just a photo and the right prompt.
How ChatGPT creates your caricature
- Analyses your photo: Facial features, expressions, and proportions.
- Applies your prompt: Defines style, background, mood, and exaggeration.
- Adds context: Combines visual cues with your chat history.
- Refines output: More detailed prompts = more personalised results.
Step-by-step: How to create an AI workplace caricature
- Log in to OpenAI: Ensure image upload is enabled.
- Upload a clear photo: Good lighting, visible face, plain background.
- Enter a detailed but safe prompt:
Example: “Create a high-quality caricature of me based on this photo. Keep my features recognisable but slightly exaggerated. Show me as a [profession] in a [setting — office, studio, cafe]. Use [style — cartoon, Pixar, hand-drawn] with vibrant colours and soft lighting.”
- Optional add-ons: Playful expressions, neutral props (laptops, books, coffee mugs).
- Review and refine: Regenerate until satisfied.
- Download and store locally: Share online only if comfortable.
Experiment with styles: hand-drawn, Pixar/3D, minimalist, bold caricatures.
Why this trend is different
Older apps only changed your facial features. ChatGPT incorporates context, creating a work identity:
- Job role and responsibilities
- Past chat topics
- Hobbies and interests
- Communication style
The result: a cartoon that feels personal — but it also reveals information about your life and work. But behind the fun lies a bigger question: how much personal and professional information are you sharing?
Your digital footprint: What you’re really sharing
To improve accuracy, users often include:
- Employer name and department
- Job responsibilities and projects
- Daily routines and tools
- Travel plans and client types
- Family references
Once posted online, these images can be:
- Downloaded or screenshotted
- Shared, reposted, or indexed
- Used for impersonation or phishing attacks
The more detail in the image, the easier it is to connect to your real identity.
Could this affect your job and company data?
Even playful AI-generated images can reveal sensitive work details: your team, projects, office setup, or clients. Sharing documents or drafts on public AI platforms may breach policies, and combining your face with work info makes phishing and impersonation easier.
Key risks include:
Even fun, stylised images can reveal:
- Workplace layouts or internal systems
- Department or team structures
- Ongoing projects
- Daily routines or tools used
- Employer or client information
Combined with your name and visual likeness, this can:
- Help scammers craft realistic phishing or impersonation attempts
- Breach non-disclosure agreements or internal IT policies
- Expose confidential company data if photos include screens, whiteboards, or documents
- Trigger compliance issues in regulated sectors (finance, healthcare, government, tech)
Oversharing daily work on AI tools adds to the risk:
- Drafting emails, reports, or contracts
- Summarising sensitive PDFs or proposals
- Refining financial forecasts or internal strategy notes
Even if no breach occurs, uploading sensitive company content to public AI platforms may violate policies and put your job at risk. If you wouldn’t post the information on LinkedIn, think twice before pasting it into an AI tool.
Takeaway: Even exaggerated caricatures can reveal real-world details. In competitive or security-sensitive industries, small information leaks can have outsised consequences.
Where your data goes
AI platforms may use submitted content to:
- Provide and maintain services
- Improve products
- Conduct research
- Share with affiliates or service providers
Pro tip: Features like chat history or memory may be on by default. Deleting content may not remove it completely.
Why the trend feels harmless
Many people love how “accurate” their caricature looks. That accuracy is not magic. It reflects information already shared over time. When the output feels personal, it creates a sense of familiarity and trust. But the AI is not discovering secrets — it is combining what you have already provided. Accuracy is built on aggregation.
Safe AI caricature tips
If you want to try the trend, keep it simple:
- Use temporary/private sessions; disable history if possible.
- Avoid real workplace photos — no screens, whiteboards, badges, or logos.
- Keep prompts general — skip employer names, projects, or client info.
- Remove metadata (location, timestamps) from images.
- Never upload sensitive work content — summarise ideas instead.
- Experiment safely with neutral props and styles.
- Store locally and share only if comfortable.
- Review platform policies periodically and delete old sessions.
The bigger picture
The ChatGPT caricature trend shows how creative and powerful AI has become. It can turn scattered details into a polished digital identity in seconds. But it also shows how easily personal and professional information can merge into a single public snapshot.
The cartoon may trend for a week. Your digital footprint — and its impact on your career — can last much longer.
GN
-
Discover3 months agoIs February 2026 really a once-in -283-years MiracleIn?
-
Football4 months agoAlgeria, Burkina Faso, Côte d’Ivoire win AFCON 2025 openers
-
Entertainment3 months agoNetflix to Livestream BTS Comeback Concert
-
Health4 months agoNMC Royal Hospital, Khalifa City, performs rare wrist salvage, restoring function for young patient
-
Health4 months agoBascom Palmer Eye Institute Abu Dhabi and Emirates Society of Ophthalmology Sign Strategic Partnership Agreement
-
Health5 months agoEmirates Society of Colorectal Surgery Concludes the 3rd International Congress Under the Leadership of Dr. Sara Al Bastaki
-
Lifestyle5 months agoSaudi Arabia Lifestyle Trends 2025: What You Need to Know About Fitness, Wellness, Healthy Eating & Self-Care Growth
-
Health5 months agoBorn Too Soon: Understanding Premature Birth and the Power of Modern NICU Care
