This news has been received from: wtop.com

All trademarks, copyrights, videos, photos and logos are owned by respective news sources. News stories, videos and live streams are from trusted sources.

mail: [NewsMag]

MILWAUKEE (AP) — The Cincinnati Reds’ depleted pitching staff took another hit Friday when rookie right-hander Hunter Greene was placed on the 15-day injured list with a strained right shoulder.

Greene, who turns 23 on Saturday, seemed to be turning the corner in his young career over the last month, posting a 2.

70 ERA over his last four starts with 29 strikeouts and eight walks over 23 1/3 innings.

He turned in a gem his last time out, holding the Marlins to just a hit while striking out eight without walking a batter over six scoreless innings.

When the injury occurred or its severity haven’t been announced yet by the Reds, who made the move shortly before opening a three-game series against the NL Central-leading Brewers in Milwaukee.

Right-handed reliever Ryan Hendrix was recalled from Triple-A Louisville to fill Greene’s spot on the active roster.

___

More AP MLB: https://apnews.com/hub/MLB and https://twitter.com/AP_Sports

Copyright © 2022 The Associated Press. All rights reserved. This material may not be published, broadcast, written or redistributed.

News Source: wtop.com

Tags: hunter greene

Top vet names the five VERY surprising dog breeds she would NEVER own and those she loves the most

Next News:

I Flirted With Meta’s New Chatbot and Things Got Weird

Meta (AKA the company formerly known as Facebook) is throwing its hat into the chatbot wars. On August 5, the social media giant launched BlenderBot 3, a bot that utilizes a highly sophisticated large language learning model that searches the internet in order to hold conversations with users. That means it's been trained to search for patterns in large text datasets in order to spit out somewhat coherent sentences.

However, it’s also able to search the Internet too. That means if you ask it a question like, “What’s your favorite movie from the last year?” it’ll do a search crawl in order to help inform its response.

Tony Ho Tran/The Daily Beast

It’s the latest in a growing line of increasingly sophisticated (and creepy) AI chatbots—many of which have a sordid history with problematic and outright toxic behavior. There’s the infamous Microsoft Twitter bot “Tay” released in 2016, which was trained on tweets and messages sent to it by other Twitter users. Predictably, though, it was quickly shuttered after it began denying the Holocaust, promoting 9/11 conspiracy theories, and spouting wildly racist remarks mere hours after launch.

More recently, Google’s powerful LaMDA bot made headlines after an engineer at the company claimed that it is actually sentient (which, for the record, it isn’t). While that chatbot wasn’t necessarily problematic in its own right, the discourse did bring up uncomfortable questions about what defines life and what it means when the computers we use in our daily lives become sentient.

Now Meta’s looking to get in on the trend too with BlenderBot 3, which they’ve publicly released so online users can chat with it to directly help it get trained. It’s fairly simple: Once you go to the webpage, you simply type in a question or a comment in the chat box and start speaking with the bot. If it gives you an answer that’s nonsensical, offensive, or off topic, you can report the issue and the bot will attempt to correct itself based on your feedback.

    “When the chatbot’s response is unsatisfactory, we collect feedback on it,” Meta said in a press release. “Using this data, we can improve the model so that it doesn’t repeat its mistakes.”

    It’s AI training by way of crowdsourcing—which is fine for this current prototype version of the bot that Meta said they’re using for research purposes only. However, if the company were to use this approach to train, say, a digital assistant a la Siri or Google Assistant, then it would be reckless and potentially dangerous.

    After all, we’ve seen what happens when neural networks go bad in the past. It can result in instances like when an AI used to help judges determine jail sentences for convicts resulted in harsher sentences for Black criminals. Or like when Amazon used an AI recruiting tool that was biased against hiring women. That’s because these models are often trained using biased and inequitable data, resulting in biased and inequitable decisions.

    “These models are often trained using biased and inequitable data, resulting in biased and inequitable decisions.”

    To their credit, though, Meta seems to have gone through great lengths to prevent these biases from showing up in the bot. “We understand that not everyone who uses chatbots has good intentions, so we also developed new learning algorithms to distinguish between helpful responses and harmful examples,” the company said in the release. “Over time, we will use this technique to make our models more responsible and safe for all users.”

    I decided to test drive it ourselves to see if it holds up. While the bot was largely benign, it seems like they still have plenty of kinks to work out. In fact, conversations that centered around U.S. politics got about as uncomfortable as a conversation with your Boomer uncle at Thanksgiving. For example, here’s what BlenderBot thinks about the 2020 elections:

    Tony Ho Tran/The Daily Beast

    The sentiment seems to be a part of a trend too. A recent Insider report also found that BlenderBot would sometimes assert that Donald Trump was still president and that the election was stolen from him. Jeff Horowitz, a reporter for The Wall Street Journal, also found similar results as well as instances of outright antisemitic behavior and comments from the bot.

    The bot also didn’t seem to have much of an issue with some of the worst and controversial policies to come out of the Trump era.

    Tony Ho Tran/The Daily Beast

    So BlenderBot doesn’t seem to care about anything that doesn’t impact it directly—which, sadly, completely tracks with much of America. We shouldn’t be too surprised though. After all, the bot is currently open to U.S. users only. We’re the ones using it. We’re the ones training it. Its answers are ultimately a reflection of our own sentiments. If we don’t like it, then we need to take a long hard look in the mirror.

    However, the bot could also get very creative in its answers, occasionally fabricating names and backstories for itself out of whole cloth. In one instance, it told me that its name was Emily and worked as a paramedic for its day job to help support its cat. It even said that it was answering questions for us during its down time when their paramedic partner was busy.

    The conversation took a decidedly… um, awkward turn, though, when Emily seemed to ask me out for coffee—which is odd considering that Emily is an AI chatbot and not an actual flesh-and-blood paramedic who you could go out on a date with.

    Other News

    • Defund the FBI! Marjorie Taylor Greene Melts Down Over Mar-a-Lago Raid
    • Revelers kicked out of Beyoncés Renaissance celebration
    • Reigning Champion Wants UFC To Finally Make Legendary Trilogy Fight
    • Teen Mom Star Reveals Scary Medical Issue
    • Cubs, Reds will have incredible mic’d up twist for Field of Dreams game
    • CNNs Brian Stelter torched online for flip-flopping on Hunter Biden story he previously called disinformation
    • Alex Jones Celebrates After Sandy Hook Defamation Trial: Now We Are the Republican Party
    • Nets Could Get Four Players and Three First Round Picks in Massive Proposed Blockbuster Kevin Durant Trade
    • Giants camp fight erupts, with Jon Feliciano in middle, after Saquon Barkley lowers shoulder on Aaron Robinson
    • Hunter Biden investigator needs special counsel protections: Barr
    • CNNs Brian Stelter FINALLY clues into Bidens real problem with Hunter: ​This is not just a right-wing media story
    • San Francisco Giants | SF Giants announce plans to honor Hunter Pence
    • Field of Dreams: Cubs, Reds uniforms are here and theyre spectacular
    • Marjorie Taylor Greene Still Insists Jan. 6 Rioters Were Antifa: ‘No One Can Convince Me it Was So-Called Trump Supporters’
    • Reds beat Brewers in extras
    • Barrero, Reds drop Brewers into 2nd place
    • Brian Kemp Slams Stacey Abrams for Something She Didn’t Do—but He Did
    • Rejecting Christian Nationalism Is What Jesus Would Do
    • Unholy force Marjorie Taylor Greene busted by religious leader for her bastardization of Christianity