Skip to content

TST: win32 paths cannot be turned into URLs by prefixing them with "file://" #4580

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 3 commits into from
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 2 additions & 0 deletions doc/source/release.rst
Original file line number Diff line number Diff line change
Expand Up @@ -39,6 +39,8 @@ pandas 0.13

**Bug Fixes**

- Fixed html tests on win32. (:issue:`4580`)

pandas 0.12
===========

Expand Down
34 changes: 30 additions & 4 deletions pandas/io/common.py
Original file line number Diff line number Diff line change
@@ -1,14 +1,25 @@
"""Common IO api utilities"""

import sys
import urlparse
import urllib2
import zipfile
from contextlib import contextmanager, closing
from StringIO import StringIO


from pandas.util import py3compat

if py3compat.PY3: # pragma: no cover
import urllib.parse as urlparse
from urllib.parse import urljoin
from urllib.request import pathname2url
from io import StringIO
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

you should import StringIO from compat. only io specific things should come from here.

else:
import urlparse
from urlparse import urljoin
from urllib import pathname2url
from StringIO import StringIO


_VALID_URLS = set(urlparse.uses_relative + urlparse.uses_netloc +
urlparse.uses_params)
_VALID_URLS.discard('')
Expand Down Expand Up @@ -68,8 +79,8 @@ def get_filepath_or_buffer(filepath_or_buffer, encoding=None):
else:
errors = 'replace'
encoding = 'utf-8'
bytes = filepath_or_buffer.read().decode(encoding, errors)
filepath_or_buffer = StringIO(bytes)
raw_bytes = filepath_or_buffer.read().decode(encoding, errors)
filepath_or_buffer = StringIO(raw_bytes)
return filepath_or_buffer, encoding
return filepath_or_buffer, None

Expand All @@ -91,6 +102,21 @@ def get_filepath_or_buffer(filepath_or_buffer, encoding=None):
return filepath_or_buffer, None


def path_to_url(path):
"""
converts an absolute native path to a FILE URL.

Parameters
----------
path : a path in native format

Returns
-------
a valid FILE URL
"""
return urljoin('file:', pathname2url(path))


# ----------------------
# Prevent double closing
if py3compat.PY3:
Expand Down
20 changes: 13 additions & 7 deletions pandas/io/tests/test_html.py
Original file line number Diff line number Diff line change
@@ -1,10 +1,16 @@
import os
import re
from cStringIO import StringIO
from unittest import TestCase
import warnings
from distutils.version import LooseVersion

from pandas.util import py3compat

if py3compat.PY3:
from io import StringIO
else:
from cStringIO import StringIO
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

ditto.


import nose
from nose.tools import assert_raises

Expand All @@ -19,14 +25,14 @@

from pandas.io.html import read_html
from pandas.io.common import urlopen
from pandas.io.common import path_to_url
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

please combine these imports


from pandas import DataFrame, MultiIndex, read_csv, Timestamp
from pandas.util.testing import (assert_frame_equal, network,
get_data_path)

from pandas.util.testing import makeCustomDataframe as mkdf


def _have_module(module_name):
try:
import_module(module_name)
Expand Down Expand Up @@ -292,7 +298,7 @@ def test_bad_url_protocol(self):
@slow
def test_file_url(self):
url = self.banklist_data
dfs = self.run_read_html('file://' + url, 'First',
dfs = self.run_read_html(path_to_url(url), 'First',
attrs={'id': 'table'})
self.assertIsInstance(dfs, list)
for df in dfs:
Expand Down Expand Up @@ -338,7 +344,7 @@ def test_multiindex_header_index_skiprows(self):
@slow
def test_regex_idempotency(self):
url = self.banklist_data
dfs = self.run_read_html('file://' + url,
dfs = self.run_read_html(path_to_url(url),
match=re.compile(re.compile('Florida')),
attrs={'id': 'table'})
self.assertIsInstance(dfs, list)
Expand Down Expand Up @@ -462,9 +468,9 @@ def test_invalid_flavor():
flavor='not a* valid**++ flaver')


def get_elements_from_url(url, element='table', base_url="file://"):
def get_elements_from_file(url, element='table'):
_skip_if_none_of(('bs4', 'html5lib'))
url = "".join([base_url, url])
url = path_to_url(url)
from bs4 import BeautifulSoup
with urlopen(url) as f:
soup = BeautifulSoup(f, features='html5lib')
Expand All @@ -476,7 +482,7 @@ def test_bs4_finds_tables():
filepath = os.path.join(DATA_PATH, "spam.html")
with warnings.catch_warnings():
warnings.filterwarnings('ignore')
assert get_elements_from_url(filepath, 'table')
assert get_elements_from_file(filepath, 'table')


def get_lxml_elements(url, element):
Expand Down