<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xml:base="http://www.softalkapple.com"  xmlns:dc="http://purl.org/dc/elements/1.1/">
<channel>
 <title>Timlynn Babitsky&#039;s blog</title>
 <link>http://www.softalkapple.com/blogs/timlynn-babitsky</link>
 <description></description>
 <language>en</language>
<item>
 <title>Archive Update - Slice Baby Slice</title>
 <link>http://www.softalkapple.com/blogs/archive-update-slice-baby-slice</link>
 <description>&lt;div class=&quot;field field-name-body field-type-text-with-summary field-label-hidden view-mode-rss&quot;&gt;&lt;div class=&quot;field-items&quot;&gt;&lt;div class=&quot;field-item even&quot; property=&quot;content:encoded&quot;&gt;&lt;p&gt;Print media pages come in all levels of complexity from a &lt;ins&gt;simple page&lt;/ins&gt; with only one large text block to a &lt;ins&gt;complex page&lt;/ins&gt; with text, tables, graphics, photos, advertisements, lists, and more. Most of the pages in the 48 issues of the Softalk magazine can be considered moderately complex, some are very complex.&lt;br /&gt;&lt;/p&gt;&lt;div class=&quot;image-right&quot;&gt;&lt;img src=&quot;/sites/default/files/images/ABBYY.jpg&quot; width=&quot;219&quot; height=&quot;419&quot; alt=&quot;ABBYY.jpg&quot; /&gt;&lt;/div&gt;
&lt;p&gt;Thanks to Peter Caylor’s monumental scanning feat we have lossless 600 dpi PDFs of all 9,300+ Softalk magazine pages. These were scanned as 2-up pages, exactly as you see them in the print copy of the magazine. In order to proceed with the &lt;strong&gt;Optical Character Recognition (OCR) phase&lt;/strong&gt; of our project, the &lt;strong&gt;2-up pages have to be split into 1-up page format&lt;/strong&gt;. This is what I’ve been working on.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;ABBYY Fine Reader 12 Professional&lt;/strong&gt; is amazing software. It can open a PDF file and through selection of various pre-processing choices, will split 2-up pages into 1-up pages. This splitting process is time-consuming (on my machine) – it takes nearly a minute a page for ABBYY to pre-process and split 2-ups. A thirty-six page issue takes a little over thirty minutes for pre-processing. But, the output is incredibly good.&lt;/p&gt;
&lt;p&gt;Because of the complexity of the pages, &lt;strong&gt;every page must be carefully “verified”&lt;/strong&gt; as having come through this page splitting process unchanged in any way other than being split into 1-ups. During the pre-processing, advertisements sometimes skew page text or cause streaking on a page. Two-ups that bleed photos across the two pages often do not get split automatically. And so I&#039;ve found (so far in Volume One) that between 5% and 11% of the pages in every issue have to be reprocessed using a multi-step manual reprocessing protocol to split, but preserve, the original pages. ABBYY Fine Reader does an amazing job on splitting the other 89-95% of the pages; just a few pages are &quot;gotchas.&quot;&lt;/p&gt;
&lt;p&gt;As of today, all the pages of all twelve issues of Volume One have been split to 1-up pages. Each page has been verified and manually reprocessed where needed. &lt;strong&gt;All issues of Volume One are now ready for the next phase -- OCR processing&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;By the end of the month, 1-up page pre-processing by ABBYY Fine Reader of all four volumes of the Softalk archive will be complete. Verifying all pages in all issues from Volume Two through Volume Four will then begin.  &lt;/p&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/div&gt;</description>
 <pubDate>Fri, 11 Jul 2014 22:35:55 +0000</pubDate>
 <dc:creator>Timlynn Babitsky</dc:creator>
 <guid isPermaLink="false">139 at http://www.softalkapple.com</guid>
 <comments>http://www.softalkapple.com/blogs/archive-update-slice-baby-slice#comments</comments>
</item>
<item>
 <title>The Future of Museums Is Open, Social, Peer-to-Peer, and Read/Write</title>
 <link>http://www.softalkapple.com/blogs/future-museums-open-social-peer-peer-and-readwrite</link>
 <description>&lt;div class=&quot;field field-name-field-tags field-type-taxonomy-term-reference field-label-hidden view-mode-rss&quot;&gt;&lt;ul class=&quot;field-items&quot;&gt;&lt;li class=&quot;field-item even&quot;&gt;&lt;a href=&quot;/tags/codewords&quot; typeof=&quot;skos:Concept&quot; property=&quot;rdfs:label skos:prefLabel&quot; datatype=&quot;&quot;&gt;#CODEWORDS&lt;/a&gt;&lt;/li&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div class=&quot;field field-name-body field-type-text-with-summary field-label-hidden view-mode-rss&quot;&gt;&lt;div class=&quot;field-items&quot;&gt;&lt;div class=&quot;field-item even&quot; property=&quot;content:encoded&quot;&gt;&lt;div class=&quot;image-right&quot;&gt;&lt;img src=&quot;/sites/default/files/images/Code_Words.jpg&quot; width=&quot;350&quot; height=&quot;207&quot; alt=&quot;Code_Words.jpg&quot; /&gt;&lt;/div&gt;
&lt;p&gt;I’ve just re-read for the third time, &lt;a href=&quot;https://medium.com/code-words-technology-and-theory-in-the-museum/a6c7430d84d1&quot;&gt;&lt;strong&gt;Michael Peter Edson’s very powerful piece&lt;/strong&gt;&lt;/a&gt; &lt;em&gt;Dark Matter&lt;/em&gt;, the first essay in &lt;a href=&quot;https://medium.com/code-words-technology-and-theory-in-the-museum/f63dabc61f47&quot;&gt;&lt;strong&gt;CODE/WORDS&lt;/strong&gt;&lt;/a&gt;, a collaborative writing project about technology in museums. &lt;/p&gt;
&lt;p&gt;Edson’s message to all of us involved with museums, libraries, and big data archives is to &lt;strong&gt;think MUCH BIGGE&lt;/strong&gt;R – way beyond tracking how many visitors come through the doors, or how many scholarly articles have come from connected researchers, or even how many page views a website has garnered.&lt;/p&gt;
&lt;p&gt;There are over 3 billion people on line today, with another 5 billion predicted to join them over the next 10 years. Museums need to understand the unprecedented opportunity they have to &lt;strong&gt;engage users to participate&lt;/strong&gt; in citizen-action science, archiving, and exploring - &lt;strong&gt;to use and share and help&lt;/strong&gt;.&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;&quot;There’s just an enormous, humongous, gigantic audience out there connected to the Internet that is starving for authenticity and good ideas—and they want to learn.&quot;&lt;/p&gt;&lt;/blockquote&gt;
&lt;p&gt;Tech innovation is leaping forward exponentially; the connected world we live in is changing almost faster than we can track. As keepers of the artifacts of human experience, museums, libraries and all related institutions need to prepare now for the &lt;strong&gt;open, democratic sharing of information&lt;/strong&gt; that the World Wide Web provides and tech advances ensure.&lt;/p&gt;
&lt;p&gt;It is no longer enough to just archive and digitize our huge collections (although that in itself is a HUGE undertaking). &lt;strong&gt;We need new platforms that allow each of us to explore those collection&lt;/strong&gt;s to gather facts, to ask questions, and to learn.&lt;a href=&quot;http://factminers.org/&quot;&gt; &lt;strong&gt;FactMiners&lt;/strong&gt;&lt;/a&gt; is one such platform and the &lt;strong&gt;Softalk Apple Project&lt;/strong&gt; provides the pilot project on which to test it.&lt;/p&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/div&gt;</description>
 <pubDate>Tue, 27 May 2014 21:51:21 +0000</pubDate>
 <dc:creator>Timlynn Babitsky</dc:creator>
 <guid isPermaLink="false">123 at http://www.softalkapple.com</guid>
 <comments>http://www.softalkapple.com/blogs/future-museums-open-social-peer-peer-and-readwrite#comments</comments>
</item>
</channel>
</rss>
