Facebook is coming for your children.
Is that framing too stark? Maybe it’s not stark enough.
Facebook recently introduced Messenger Kids, a version of their Messenger app designed for six to twelve year olds. Antigone Davis, Facebook’s Public Policy Director and Global Head of Safety, wrote a blog post introducing Messenger Kids and assuring parents the app is safe for kids.
“We created an advisory board of experts,” Davis informs us. “With them, we are considering important questions like: Is there a ‘right age’ to introduce kids to the digital world? Is technology good for kids, or is it having adverse affects on their social skills and health? And perhaps most pressing of all: do we know the long-term effects of screen time?”
The very next line of Davis’s post reads, “Today we’re rolling out our US preview of Messenger Kids.”
Translation: We hired a bunch of people to ask important questions. We have no idea what the answers may be, but we built this app anyway.
Davis doesn’t even attempt to fudge an answer to those questions. She raises them and never comes back to them again. In fact, she explicitly acknowledges “we know there are still a lot of unanswered questions about the impact of specific technologies on children’s development.” But you know, whatever.
Naturally, we’re presented with statistics about the rates at which children under 13 use the Internet, Internet-enabled devices, and social media. It’s a case from presumed inevitability. Kids are going to be online whether you like it or not, so they might as well use our product. More about this in a moment.
We’re also told that parents are anxious about their kid’s safety online. Chiefly, this amounts to concerns about privacy or online predators. Valid concerns, of course, and Facebook promises to give parents control over their kids online activity. However, safety, in this sense, is not the only concern we should have. A perfectly safe technology may nonetheless have detrimental consequences for our intellectual, moral, and emotional well-being and for the well-being of society when the technology’s effects are widely dispersed.
Finally, we’re given five principles Facebook and its advisory board developed in order to guide the development of their suite of products for children. These are largely meaningless sentences composed of platitudes and buzzwords.
Let’s not forget that this is the same company that “offered advertisers the opportunity to target 6.4 million younger users, some only 14 years old, during moments of psychological vulnerability, such as when they felt ‘worthless,’ ‘insecure,’ ‘stressed,’ ‘defeated,’ ‘anxious,’ and like a ‘failure.'”
Facebook doesn’t care about your children. Facebook cares about your children’s data. As Wired reported, “The company will collect the content of children’s messages, photos they send, what features they use on the app, and information about the device they use.”
There are no ads on Messenger Kids the company is quick to point out. “For now,” I’m tempted to add. Barriers of this sort tend to erode over time. Moreover, even if the barrier holds, an end game remains.
“If they are weaned on Google and Facebook,” Jeffrey Chester, executive director for the Center of Digital Democracy, warns, “you have socialized them to use your service when they become an adult. On the one hand it’s diabolical and on the other hand it’s how corporations work.”
Facebook’s interest in producing an app for children appears to be a part of a larger trend. “Tech companies have made a much more aggressive push into targeting younger users,” the same Wired article noted, “a strategy that began in earnest in 2015 when Google launched YouTube Kids, which includes advertising.”
In truth, I think this is about more than just Facebook. It’s about thinking more carefully about how technology shapes our children and their experience. It is about refusing the rhetoric of inevitability and assuming responsibility.
Look, what if there is no safe way for seven-year-olds to use social media or even the Internet and Internet-enabled devices? I realize this may sound like head-in-the-ground overreaction, and maybe it is, but perhaps it’s worth contemplating the question.
I also realize I’m treading on sensitive ground here, and I want to proceed with care. The last thing over-worked, under-supported parents need is something more to feel guilty about. Let’s forget the guilt. We’re all trying to do our best. Let’s just think together about this stuff.
As adults, we’ve barely got a handle on the digital world. We know devices and apps and platforms are designed to capture and hold attention in a manner that is intellectually and emotionally unhealthy. We know that these design choices are not made with the user’s best interest in mind. We are only now beginning to recognize the personal and social costs of our uncritical embrace of constant connectivity and social media. How eager should we be to usher our children in to this reality?
The reality is upon them whether we like it or not, someone might counter. Maybe, but I don’t quite buy it. Even if it is, the degree to which this is the case will certainly vary based in large part upon the choices parents make and their resolve.
Part of our problem is that we think too narrowly about technology, almost always in terms of functionality and safety. With regards to children, this amounts to safeguarding against offensive content, against exploitation, and against would-be predators. Again, these are valid concerns, but they do not exhaust the range of questions we should be asking about how children relate to digital media and devices.
To be clear, this is not only about preventing “bad things” from happening. It is also a question of the good we want to pursue.
Our disordered relationship with technology is often a product of treating technology as an end rather than a means. Our default setting is to uncritically adopt and ask questions later if at all. We need, instead, to clearly discern the ends we want to pursue and evaluate technology accordingly, especially when it comes to our children because in this, as in so much else, they depend on us.
Some time ago, I put together a list of 41 questions to guide our thinking about the ethical dimensions of technology. These questions are a useful way of examining not only the technology we use but also the technology to which we introduce our children.
What ideals inform the choices we make when we raise children? What sort of person do we hope they will become? What habits do we desire for them cultivate? How do we want them to experience time and place? How do we hope they will perceive themselves? These are just a few of the questions we should be asking.
Your answers to these questions may not be mine or your neighbor’s, of course. The point is not that we should share these ideals, but that we recognize that the realization of these ideals, whatever they may be for you and for me, will depend, in greater measure than most of us realize, on the tools we put in our children’s hands. All that I’m advocating is that we think hard about this and proceed with great care and great courage. Great care because the stakes are high; great courage because merely by our determination to think critically about these matters we will be setting ourselves against powerful and pervasive forces.