T-Mobile is once again being sued by Washington state over the 2021 data breach which exposed sensitive information for over 79 million people, The Verge reports. The lawsuit filed on Monday alleges that T-Mobile had been aware of various security loopholes in its systems for years but didn’t take any action. As a result, a hacker managed to breach T-Mobile in March 2021 and was undetected until August of the same year when an “anonymous cybersecurity threat intelligence firm” told T-Mobile what was happening.
Beyond alleging that T-Mobile knew about these flaws and took inadequate action to fix them, Washington State Attorney General Bob Ferguson also claims T-Mobile’s notifications to customers affected by the breach were inadequate and misleading. The text messages were brief and didn’t reveal the full scope of the breach, only telling customers that debit and credit card information wasn’t exposed while failing to mention their social security numbers and other personally identifiable information were compromised.
The breach’s victims included two million Washington residents. Information from T-Mobile’s databases was later on the dark web for sale to the highest bidder. T-Mobile even supposedly hired a third party to buy exclusive access to the data.
In more than one sense, this isn’t T-Mobile’s first rodeo. The company was already sued by AG Ferguson over a decade ago over “deceptive” ads. It has also been the target of a breach since 2021 — specifically 2024 “Salt Typhoon” attacks on commercial telecommunications companies. T-Mobile claims that its systems and data weren’t impacted significantly.
This article originally appeared on Engadget at https://www.engadget.com/cybersecurity/t-mobile-is-under-fire-again-over-its-2021-data-breach-143007400.html?src=rss
NASA will analyze and explore two different landing options for its Mars Sample Return program, though it will take almost two years to do so and is expected to announce its decision in late 2026. The agency had to temporarily hit pause on the program after an independent review found that it could cost between $8 billion and $11 billion, which is way above budget.
The first method NASA is evaluating is called the “sky crane,” in which a vehicle will head to Mars, get close to the surface with the help of a parachute, pick up the samples the Perseverance rover had collected using cables or other mechanisms and then fly away. NASA previously used this method to place the Curiosity and Perseverance rovers on the planet.
Meanwhile, the second option requires the help of commercial space companies. Last year, the agency asked SpaceX, Blue Origin, Lockheed Martin and other companies to submit proposals on how to get the collected Martian samples back to Earth. Whichever option the agency chooses will carry a smaller version of the Mars Ascent Vehicle than originally planned. The Mars Ascent Vehicle is a lightweight rocket that will take the samples from the planet’s surface into Martian orbit. It will also have to be capable of transporting a container that can fit 30 sample tubes. Once the sample container is in orbit, a European Space Agency orbiter will capture it and bring it back home.
Early last year, NASA’s Jet Propulsion Laboratory had to lay off 530 employees and cut off 100 contract workers mainly due to budget issues related to this mission. NASA requested $950 million for the program, but only $300 million was allocated for it. The independent review that found that the mission would cost above budget also found that it might not be able to bring the samples back to Earth by 2040. According to a previous report by The Washington Post, the US government found the return date “unacceptable.”
In a teleconference, NASA administrator Bill Nelson revealed either of the two methods the agency is now considering would cost a lot less than what it would originally spend. The sky crane would reportedly cost NASA between $6.6 billion and $7.7 billion, while working with a private space company would cost between $5.8 billion and $7.1 billion. Either option would also be able to retrieve the samples and bring them back sometime between 2035 and 2039. Scientists believe the samples Perseverance has been collecting could help us determine whether there was life on Mars and whether its soil contains chemicals and substances that could be harmful to future human spacefarers.
This article originally appeared on Engadget at https://www.engadget.com/science/space/nasa-will-decide-how-to-bring-soil-samples-back-from-mars-in-2026-141519710.html?src=rss
How can numerical user metrics, such as “3 visits in the past week,” be transformed into a personalized assessment of whether this behavior is typical or unusual for the user?
Cover, image by Author
In almost any digital product, analysts often face the challenge of building a digital customer profile — a set of parameters that describe the customer’s state and behavior in one way or another.
What are the potential applications of these metrics?
Gaining insights into user behavior
Leveraging as features in ML models
Developing business rules for personalized offers
A simple example is e-commerce, with metrics like those listed in the table below.
Image by Author
These metrics are used everywhere, but what is the problem with them? They don’t take into account the specific user’s history or the dynamics of this particular metric. Is $200 spend a lot for user 1? It’s unclear. Yet this distinction significantly influences the business decision we make next.
Even within the context of a single user, $200 can have a different meaning for the business depending on the user’s stage in their lifecycle with the product. $200 spent during the user onboarding, pick activity and re-activation are different.
User journey, image by Author
We’d like to have some normalized metrics to be able compare them between users. Something like this:
Image by Author
So how can we move from a numerical description of customer behavior to a more characteristic representation? For instance, how can the fact that “a customer hasn’t made a transaction for 7 days” be translated into an individualized assessment of whether this is unusual or typical for that specific customer? The challenge is to achieve this without losing interpretability, preserving business relevance, and avoiding black-box models.
A simple approach is to analyze the distribution of the metric and determine the probability of observing the current result (i.e., calculate the p-value). This helps us understand how extreme the value is compared to the user’s history.
Normal distribution, image by Author
But what’s the challenge here? In most cases, the distribution of metrics is not normal, making p-value calculations more complex.
The random metric probably would have a distribution similar to this one:
PDF, image by Author
We can apply some small trick and to transform Probability Density Function to Cumulative Distribution Function (CDF). Calculating p-value in this case is much easier.
CDF, image by Author
So, we simply need to reconstruct the CDF from the user’s metric, which can be done efficiently using splines. Let’s create a toy example.
Imagine you are an e-commerce platform aiming to personalize your email campaigns based on user activity from the past week. If a user has been less active compared to previous weeks, you plan to send them a discount offer.
You’ve gathered user statistics and noticed the following for a user named John:
John visited the platform for the first time 15 days ago.
During the first 7 days (days 1–7), he made 9 visits.
During the next 7 days (days 2–8), he made 8 visits.
Totally we have 9 values.
Now, you want to evaluate how extreme the most recent value is compared to the previous ones.
Not bad. But we observe a small problem between red dots — the CDF must be monotonically increasing. Let’s fix this with Piecewise Cubic Hermite Interpolating Polynomial.
CDF with Piecewise Cubic Hermite Interpolating, image by Author
Alright, now it’s perfect.
To calculate p-value for our current observation (6 visits during the last week) we need to calculate the surface of filled area.
Critical area, image by Author
To do so let’s create a simple function calculate_p_value:
def calculate_p_value(x): if x < values.min(): return 0 elif x > values.max(): return 1 else: return spline_monotonic(x)
p_value = calculate_p_value(num_visits_last_week) print(f"Probability of getting less than {num_visits_last_week} equals: {p_value}")
Probability of getting less than 6 equals: 0.375
So the probability is quite high (we may compare it to a threshold 0.1 for instance) and we decide not to send the discount to John. Same calculation we need to do for all the users.
Conclusion
In this article, we have seen how transforming raw user metrics into personalized assessments can lead to more effective business decisions. By using statistical methods like CDF and spline interpolation, we can better understand the context behind user actions and provide personalized, relevant offers that are informed by data.
We use cookies on our website to give you the most relevant experience by remembering your preferences and repeat visits. By clicking “Accept”, you consent to the use of ALL the cookies.
This website uses cookies to improve your experience while you navigate through the website. Out of these, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may affect your browsing experience.
Necessary cookies are absolutely essential for the website to function properly. These cookies ensure basic functionalities and security features of the website, anonymously.
Cookie
Duration
Description
cookielawinfo-checkbox-analytics
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Analytics".
cookielawinfo-checkbox-functional
11 months
The cookie is set by GDPR cookie consent to record the user consent for the cookies in the category "Functional".
cookielawinfo-checkbox-necessary
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookies is used to store the user consent for the cookies in the category "Necessary".
cookielawinfo-checkbox-others
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Other.
cookielawinfo-checkbox-performance
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Performance".
viewed_cookie_policy
11 months
The cookie is set by the GDPR Cookie Consent plugin and is used to store whether or not user has consented to the use of cookies. It does not store any personal data.
Functional cookies help to perform certain functionalities like sharing the content of the website on social media platforms, collect feedbacks, and other third-party features.
Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.
Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics the number of visitors, bounce rate, traffic source, etc.
Advertisement cookies are used to provide visitors with relevant ads and marketing campaigns. These cookies track visitors across websites and collect information to provide customized ads.