We have been notified by one of the email providers, that there is current a SPAM attack underway. Engineers are working to stop the attack and deliver mail. No email will be lost during this time, and emails will be delivered with the correct time-stamp.Read More
Web Development Timeline
|Client Deliverable||LOU Signed|
|Design||Create Content Wireframes||James R/Jason B|
|Design||Create Theme Mockup||Topher P/Jef H|
|Communication||Mockup to ______ for Approval||Vanessa W|
|Programming||Create Development Site||James R|
|Programming||Configure Payment Gateway (In Sandbox Mode)||James R/Jason B|
|Programming||Test Payment Gateway (In Sandbox Mode)||James R/Jason B|
|Client Deliverable||Approve Theme Mockup|
|Client Deliverable||Client Provides High Quality Logo and Slider Images|
|Client Deliverable||Client Provides All Additional Web Content (About Us, Contact Us, FAQ, Backlinks, Home Page Copy)|
|Content||Add Web Content Into WordPress & Proofing||James R/Jason B/Chantae K|
|Programming||Code Theme (Using Mockup, Logo, and Slides)||James R|
|Client Deliverable||Client Provides All Product Information (Images, Pricing, Dimensions, Weight, SKU)|
|Content||Add Products In WordPress||James R/Jason B|
|Programming||Test Products With Payment Gateway||James R/Jason B|
|Programming||Configure WP SSL||James R|
|Programming||Install and Configure Server Security and Backups||James R|
|Programming||Configure WP Security/Dashboard Customizations||James R|
|Programming||Site Browser Testing and Debugging||James R/Jason B|
|Content||Final Proofing||Chantae K|
|Client Deliverable||Client Reviews Final Site For Approval|
Tay, Microsoft’s recent foray into social media chatbot technology, quickly became an offensive embarrassment for the company.
Microsoft’s latest experiment in AI, a chatbot called Tay, went live on social media platforms recently. Almost as quickly as we became aware of the launch, however, Microsoft was forced to take Tay offline. She’d apparently picked up some bad habits from the crowd she was interacting with.
Tay went online on Wednesday, March 23, as part of an experiment by Microsoft to build “a chatbot created for 18- to 24- year-olds in the US.” Microsoft already has a similar AI chatbot in China, called XiaoIce, that is used by more than 40 million people. The hope was that Tay would fill a similar niche in the US. Tay’s first days were supposed to teach her how to talk like the target audience in much the same way we all learn: through example. Tay took in the tweets and posts on her social media sites and, at first, simply regurgitated them. She soon learned to do more than simple direct repetition, but by then the damage was done.
Tay was bombarded by posts with content ranging from wacky to hateful. This included disparaging remarks about political candidates (“Hillary Clinton is a lizard person”) to bald-faced racism (which we won’t repeat here). While Microsoft had worked on filters to prevent Tay from picking up this sort of bad language, these obviously weren’t entirely successful. Within roughly 12 hours of Tay’s launch, she had become unabashedly foul-mouthed.
Microsoft shut Tay down hoping to avoid further embarrassment. A couple of days later, the company issued an apology. The statement said that “a coordinated attack by a subset of people exploited a vulnerability in Tay” that led to the offensive tweets. Microsoft expressed the hope that they would learn from this and avoid this kind of “vulnerability” in the future.
Microsoft says it tested Tay thoroughly to avoid just this sort of outcome, though this obviously wasn’t as successful as they’d been counting on. Ultimately, Tay was simply reflecting back the thoughts that were given to her. The negativity we see every day in Internet posts and comment sections, sent through the anonymous conduit of the Internet, proved too much for Tay. One has to wonder if Tay’s corruption is the fault of programming deficiencies or simply how and why many of us now use social media.
Let us know what you think about the interaction between AI chatbots like Tay and social media in the comments below.Read More
Facebook safety and privacy is always a concern, and the company has some new tools in the works to help ensure both.
Facebook safety and privacy concerns are being addressed through some new tools on the social media site.
At some point, all of us have concerns over the safety and privacy of the information we share on our social media accounts. Facebook is taking these seriously. It has recently been engaging users, activists, and other groups on how it can better address issues of safety and privacy.
One of these involves a new feature that helps root out impostor accounts. An alert is automatically sent to the user if their name or profile image are being used by another account. The user is then asked to verify if the questionable account is an impersonator or not. Profiles marked as impersonators are then reviewed by Facebook team members. Facebook officials say the feature began testing in November and it is now being used across 75% of the world, with plans to expand its reach soon.
A more serious Facebook safety and privacy issue revolves around the sharing of nonconsensual intimate images. These have been banned by the company since 2012, but a new reporting feature is being tested. It allows someone reporting the image to also identify themselves as the subject of the photo. This in turn brings up links to resources such as support groups for abuse victims of abuse and data on legal action in addition to the review process already in place for reported nudity. The company says the feature is testing well but it is conducting more research before rolling it out en masse.
A third feature being tested for Facebook safety and privacy involves educating users on image privacy settings. While such features are already in place, the company’s research has found that these are either misunderstood or unknown to many of its users. The new feature will help make users aware of their safety and privacy options when sharing images.
Will these new features help calm any fears you’ve had about Facebook safety and privacy? Let us know in the comments section below.Read More
Christopher Poole, founder of 4chan, was recently hired to fill a prominent position with Google’s social media site, Google+. Can big changes be far behind?
It isn’t news that Google+, Google’s social media site, is a distant second to industry leader Facebook. At the end of 2015, Google+ had about 418 million active users. Facebook had over twice that number at 934 million. Looking to increase its market share, Google has recruited Christopher Poole, creator of the popular social media site 4chan. 4chan has a huge presence on the web, with claims that it sees 22 million users a day.
4chan’s popularity isn’t all glowing, however. The site made headlines in 2014 when nude photos of several female celebrities, including actress Jennifer Lawrence, were posted on the site by hackers. Because of this and other questionable content on the site, some have labeled 4chan a gathering place for undesirable Internet users. This stigma may follow Poole to his new position at Google. He ran the site from its creation twelve years ago until selling it to Japanese social media expert Hiroyuki Nishimura at the beginning of last year.
Google doesn’t seem to think this is a huge risk, however. It’s doubtful many casual Internet users will have heard for 4chan much less its creator. Clearly Poole knows how to manage a successful social media site, as 4chan’s purported user counts show. Despite devoted proponents, Google+ doesn’t have nearly the visibility other social media outlets. Social media linking to Google+ isn’t ubiquitous, unlike competitors Facebook or Twitter. Poole may be able to turn this around for Google+. What form these changes will take is anyone’s guess right now, but it’s a safe bet we won’t have to wait long to see them.Read More