Free tool to automate social media updates from website

If you ever published any new post on to your blog / website, and are using rss feed, your website adds newly published post to rss xml based on date and time of publishing and this same URL remains forever with the same timestamp. Now, if you are using automated feed burners or email services or social media tools which takes this rss xml file and sends the contents to intended audience, your post will reach to your followers only once when its actually published.

But, isn’t it would be great for us to have some kind of tool which automatically scan’s URL of your website, picks one randomly and emails it to your followers or update to your social media platform. Following description does exacly the same by taking benefit of tools like Google RSS Feed Subscriber, to generate rss xml using predefined set of URL’s already copied into one text file and taking only few of the URL’s by shuffling and creating final rss compatible xml with latest timestamp, so automated tools will consider this xml as if your post has been published recently.

$ vim
echo "<?xml version=\"1.0\" encoding=\"UTF-8\" ? >
<rss version=\"2.0\">
  <description>Linux, Android and Opensource software knowledge platform</description>" > custom-rss.xml
while read line
TZ=`date +"%z"`
TIME=`date +"%H:%M:%S"`
WEEKDAY=`date +"%a"`
DATE=`date +"%d"`
MONTH=`date +"%b"`
YEAR=`date +"%Y"`
echo "<item>" >> custom-rss.xml
echo "<title> Lynxbee $TIME </title>" >> custom-rss.xml
echo "<link>$line</link>" >> custom-rss.xml
echo "<pubDate>FINAL</pubDate>" >> custom-rss.xml
echo "<description&gt; Linux, Android and Opensource software $TIME </description>" >> custom-rss.xml
echo "</item>" >> custom-rss.xml
sleep 1
echo "</channel>
</rss>" >> custom-rss.xml

Now, create only_uniq_urls.txt file which actually contents list of total URL’s of your website, from subset of which you will create a rss xml.

$ vim only_uniq_urls.txt
Url for RSS

Now, lets generate the xml by running, script as,

$ bash

Once, you run above script, the line “shuf -n 3 $WEBSITE_URL_FILE > $SHUFFLED_URL_FILE” creates another text file shuffled_urls.txt with only 3 URLs shuffled from only_uniq_urls.txt text file and script goes on reading shuffled_urls.txt and creates an rss compatible custom-rss.xml XML file as below,

<?xml version="1.0" encoding="UTF-8" ?>
<rss version="2.0">
  <description>Linux, Android and Opensource software knowledge platform</description>
<title> Lynxbee 21:19:34 </title>
<pubDate>Wed, 27 Feb 2019 21:19:34 +0530</pubDate>
<description> Linux, Android and Opensource software 21:19:34 </description>
<title> Lynxbee 21:19:35 </title>
<pubDate>Wed, 27 Feb 2019 21:19:35 +0530</pubDate>
<description> Linux, Android and Opensource software 21:19:35 </description>
<title> Lynxbee 21:19:36 </title>
<pubDate>Wed, 27 Feb 2019 21:19:36 +0530</pubDate>
<description> Linux, Android and Opensource software 21:19:36 </description>
Anonymous Instagram story viewer: h... x
Anonymous Instagram story viewer: how to use it and download Instgram posts and stories

Leave a Comment

Sign up to our newsletter!

ADB Commands Android Android Applications Android Build system Application Libraries Application Stack / User Interface Bash / Shell Scripts Bluetooth driver Cloud Technologies Commands and Packages Compilation Content Management System Core Kernel C Programs Development, Debugging and Performance Tools Development and Build Development Environment Setup Django & REST Api Errors & Failures Git Hardware Platforms HTML JAVA Programs Linux, OS Concepts and Networking Linux Device Drivers Linux Host, Ubuntu, SysAdmin Linux Kernel Linux Networking Middleware Libraries, HAL Multimedia NDK / Middleware / HAL Operating System Concepts PHP Programming Languages Scripting and Automation Search Engine Optimisation ( SEO ) Social Media Source Code Management ( SCM ) System Administration, Security Testing and Debugging Uncategorized Web Design and Development Website Hosting Wordpress Yocto Embedded Linux