Apple's "Siri: Learn From This App" feature is enabled by default for every app on your device.
Siri is analyzing your interactions with apps, including banking, email, and messaging apps, unless you manually opt out one app at a time.
To turn it off:
1. Go to Settings.
2. Open Apps.
3. Select an app, then go to Siri settings.
4. Disable "Learn From This App".
There's NO global toggle.
Hiring managers now can can post job opportunities on https://cupid.careers, which will be published by @thecupid to the Fediverse
This is the first and main paid feature of the service. I'm still figuring out the pricing, but at the moment I'm putting it at $89/month, with a 7-day free trial, and you can cancel any time.
To help bootstrap it, there is a 2025 deal: $20.25/month for the first 45 companies - DM me please for promotion code.
Please boost for reach. Let's get #fedihired
Recently @siderea wrote a fantastic thread about social homogeneity, moderation, the design of social platforms and what they could be. They covered a lot of ground and I can’t respond to it all so I’ll just pick some highlights
I cannot tell you how many conversations I have seen about the topic of “moderation” and how necessary it is in which nobody has ever bothered to set down what exactly it is that they think a moderator is supposed to accomplish.
I mean, it’s all of them. I’ve been on the internet since the 1980s, and I have never seen anyone stop and actually talk about what they thought moderators were trying to do or should try to do.
That sounds easy. I’ll take a shot at that, below.
Also they draw a parallel between designing buildings and designing social platforms:
Why should our societies tolerate the existence of *irresponsibly* designed and operated social media platforms, that increase violence and other antisocial behavior?
Primarily buildings are built to be used, and as such they are tools, and we judge them, as we do all tools, by how fit they are for their purpose, whatever that might be.
And the purposes of buildings are to afford various ways of people interacting or avoiding interacting.
So architects think a lot about that. It’s a whole thing.
Those who put together social media platforms need to think about the same sort of thing.
Preach!
The upshot is that we can do better than what we have in the past. We can go beyond the bare minimum of “delete the spam, ban the nazis” moderation. When we build social software the features it has will determine what kind of moderation is possible, what kind of interactions people will have. We should be intentional about that.
I’d like to share some of my ideas for how we can do that but first, let’s get the basics covered:
What I think a moderator is supposed to accomplishObviously every online space is different and has it’s own values and priorities. What follows is what I consider to be the minimum necessary to avoid devolving into 4chan as soon as the normies arrive.
The goal of moderators is to create a positive, inclusive, and constructive online community where users feel comfortable engaging in discussions and sharing their thoughts and ideas. To that end, their responsibilities include:
Ok, cool. But:
We can and should accomplish moreWhen we think about moderation tools for a platform that serves millions of people, we are shaping the nature of social interactions on a grand scale. As we engineer these virtual societies, the question we need to ask ourselves is, “What is the nature of the society we want to create?” and within that, “What do we want moderation to accomplish that supports that nature?” and eventually “What software features do moderators need to do their work?”
The nature of the societyWe want to create an ideal society where everyone is safe, respected, empowered, entertained and encouraged to grow and find meaning according to their individual free choices. Members of this online society contribute meaningfully and positively to the rest of society, support the actualisation of human rights for all and work to help democracy to live up to it’s promise.
Remember the 1990s, when the internet hadn’t been corrupted yet? Yeah. I do.
What we want moderation to accomplish to maintain this ideal societyDefining the Role of Moderation
Moderation should not be a passive, reactive role. Instead, it should be proactive, shaping the community’s social dynamics intentionally. The first step towards this is defining what our platforms aim to achieve. Do we want a space for free and open discussions, a supportive community, or a platform for specific interests? This vision will shape the guidelines we develop, the tools we use, and the strategies we implement.
Developing Clear Guidelines and Empowering Moderators
Once we have our vision, we need to create a set of rules that align with this vision. These guidelines should be clear, easily accessible, and comprehensive. Moreover, we need to empower our moderators with the right tools and authority to enforce these guidelines. This can include features for deleting posts, banning users, or moving discussions.
Investing in Technology
Incorporating technology is crucial in supporting our moderators. Automated moderation tools can detect and remove inappropriate content, while algorithms can promote high-quality posts. Technology can also help in combating challenges like trolls who use new IP addresses to create accounts. Techniques like browser fingerprinting can identify users regardless of their IP, and restrictions on new accounts can deter trolls.
Addressing Complex Issues
Online communities also need to grapple with complex issues such as the formation of high-control groups, disinformation propagation, social isolation, and internet addiction. Tackling these problems requires more advanced tools and strategies:
Continuous Evaluation and User Education
Finally, moderation should be an ongoing process of improvement and adaptation. We need to regularly review and update our strategies based on their effectiveness and changing conditions. Additionally, we need to educate our users about these issues and how to report them. An informed user base can greatly aid in maintaining a healthy community.
In conclusion, moderation in online communities is not just about maintaining order but about intentionally shaping the dynamics of these spaces. As we navigate the digital age, we must recognize the power and responsibility we hold in engineering these virtual societies, and use it to create healthier, safer, and more inclusive communities.
https://join.piefed.social/2024/03/07/moderation-the-design-of-social-platforms/
Society is easier than saying, "The great mythos of ideological thought that's brought about in the vat of cultural, spiritual and moral ideology."
I mean, you are a part of (and influence society) whether you want to believe it or not. Right now you have less influence because the megaphones have been moved back to the few to blast to the masses, but as with the beginning of the printing press and the radio, these megaphones can make their way back to the people.
Just like the Radio and Printing Press, control, marketing and state actors have turned the Internet into a one directional medium devoid of discussion. But these tables do turn eventually. It requires some degree of vigilance. I'm curious what new technology will cause the next big shift.
In the past I've worked on local campaigns, I've gone to city council meetings about things I cared about (I once got up and presented on transportation options in Wellington) .. and once you start going and seeing what others say and present, it helps you make better presentations too. I haven't in a long while though, but it's something anyone can do with some time.
To quote Robert Barnes, "The greatest trick the devil ever pulled was convincing the world he didn't exist. The greatest trick the State even pulled was convincing the people you couldn't resist."
GNU social JP is a social network, courtesy of GNU social JP管理人. It runs on GNU social, version 2.0.2-dev, available under the GNU Affero General Public License.
All GNU social JP content and data are available under the Creative Commons Attribution 3.0 license.