Posts

Showing posts from March, 2016

RSS feeds to your webpage [part 2]

Have a look at your management settings in https://analytics.google.com/ on the Admin tab. This process is going slow, but it helps to know why blank ads are shown on your website: 1. Google does preliminary checks on your website to ensure it passes their checks for conformity, etc. 2. Your website pages includes the ad code generated by Adsense . After you have passed these steps and you notice your Ad Unit status being "New", you need to be patient. Google could take a few days to approve your application. You will receive a notification via email on outcome. At the same time, if approved, you website will start showing ads. That then means the cash can start rolling in.... Yippee!! Happy income generation peoples. www.silvafox.co.za

RSS feeds to your webpage [part 1]

So you would like to add some dynamism to your static html pages. Well, ensure your hosted website service provider supports php scripting, which will be strange if they do not. Download the rss2html file where you will apply your changes to. Download from this location. Unzip this file and add it to the root of your website. Here are your checkpoints: Ensure your feed is reachable via following link: http:// <blogname> .blogspot.com/feeds/posts/default?alt=rss. Ensure you replace <blogname> with your blog name Modify your rss2html.php file as below: Configuration of RSS2html Unzip the folder.  Open the rss2html-docs.txt file in Notepad++ or other plain text editor and read the instructions.  Actually it's very simple to implement by modifying a few lines of code in the rss2html.php file: Tell the script where to find your RSS Feed:   $XMLfilename = "http:// your_blog .blogspot.com/feeds/posts/default?alt=rss"; Tell the script where to f

"Bigness" data - some thoughts about the big data journey

The " Bigness " data. Ideas of Big Data If you are reading this blog, you are likely to be a nerd. Everywhere - Yes, data is everywhere, and the volumes are exponentially increasing by day. There exists a saying: "Where there is a will there is relatives". Are your measures trustworthy? Ensure their integrity, but at the same time give attention to the training needed to its users. Data Detectives - The cool part for me of data science, is the detecting fraud part. Fraudsters love lots and lots of data. The more the merrier. Its also easier to hide abnormalities when lots of data are present. Therefore, if you find your organization at the stage where you are IYO consuming and processing big data, then its time you consider employing a data detective. Machine Learning - This will only be useful if you actually start using what your system is learning about trends, etc OODA Loop - What came first? The egg or the chicken. This scenario gives ri

High speed data loading with SSIS

Image
Tasks or Containers should be run in parallel. The below Container A illustrates the parallelism we want to achieve. Container B illustrates sequential workflow: In the package property, you want to play with the property MaxConcurrentExecutables . With it, you can configure how many executables to run in parallel. Other things that are very important to keep in mind are : Reduce the number of columns - Yes, ensure to reference columns that you actually need. Lots of wasted space is saved, by following this golden rule Reduce the number of rows - If you are loading from a flat file, split the large files where possible Reduce column width - Here data type is VERY important. Overkill in data type usage is a common problem. The smaller the datatype (ensure you are not truncating important field information :-) ), the more efficient your processing will be Retrieve data using sp_executesql - When you are retrieving data, you may find you can directly using the Tabl

Importing a database dump in MySQL

The below command is one that you need to remember: mysql -u < user > -p < db_backup. dump Just fired up the windows command line, and ensured my local MySQL server is running. I then just now ran the above command to do the import. Remember, that you may specify the exact path to the dump file as well, if it does not reside in the working directory of MySQL. Also note that the .dump file could have also been originally exported as a .sql, in which case the process will be the same. This import took a 400 meg database dump about 6 minutes to load on my system. If you are running a SSD setup, expect much faster load times :-) Happy importing and exporting people www.silvafox.co.za

Moving Data [part 2]

So we've put off diving into the world of Apache Spark, purely because we were used to the other tools in industry. One could say, we were in a comfort zone with, for the most part, vendor specific load functions like LOAD DATA INFILE (MySQL), etc. We could move data, and we weren't interested in looking into the new ways of doing things. However, since we've entered the world of transformations and actions with Apache Spark, it doesn't seem like we will be looking back. It's a skill set that is rare at this stage in industry, and, it's a sought after skill set. Come talk to us regarding moving terabytes and terabytes of data in the most efficient way possible. www.silvafox.co.za

Moving Data [part 1]

Recent data transformation needs has driven us to re-look at the types of offerings we give our clients. Don't get me wrong, we love building reports. However, what businesses want is intelligence from their valuable data. This is a challenge still today. Stay tuned for how we used Apache Spark to move data in a way no other tool could do before it. www.silvafox.co.za