File: README.md

package info (click to toggle)
azure-data-lake-store-python 1.0.1-1
  • links: PTS, VCS
  • area: main
  • in suites: sid
  • size: 31,952 kB
  • sloc: python: 4,332; makefile: 192
file content (36 lines) | stat: -rw-r--r-- 1,770 bytes parent folder | download | duplicates (4)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
To run the test suite against the published package:

    py.test -x -vvv --doctest-modules --pyargs azure.datalake.store tests

To run the test suite against a local build:
    
    python setup.py develop
    
    py.test -x -vvv --doctest-modules --pyargs azure.datalake.store tests
    
This test suite uses [VCR.py](https://github.com/kevin1024/vcrpy) to record the
responses from Azure. Borrowing from VCR's
[usage](https://vcrpy.readthedocs.io/en/latest/usage.html#record-modes), this
test suite has four recording modes: `once`, `new_episodes`, `none`, `all`. The
recording mode can be changed using the `RECORD_MODE` environment variable when
invoking the test suite (defaults to `none`).

To record responses for a new test without updating previous recordings:

    RECORD_MODE=once py.test -x -vvv --doctest-modules --pyargs azure-datalake-store tests

To record responses for all tests even if previous recordings exist:

    RECORD_MODE=all py.test -x -vvv --doctest-modules --pyargs azure-datalake-store tests

When recording new responses, you will need valid Azure credentials. The following
environment variables should be defined:

* `azure_data_lake_store_name` : The data store account name, without any suffix like azuredatalakestore.net
* `azure_subscription_id`   : Subscription ID for ADLS account
* `azure_resource_group_name` : Resource group for adls account.
* `azure_service_principal` : Service principal of app with owner access to account.
* `azure_service_principal_secret`: Service principal secret with owner access to account.
* `AZURE_ACL_TEST_APPID ` : Service principal of app with access to account.

Optionally, you may need to define `azure_username`, `azure_password`, `azure_tenant_id` or `azure_data_lake_store_url_suffix`.