Replies: 3 comments
-
trimmed by sed |
Beta Was this translation helpful? Give feedback.
0 replies
-
Not sure. Can Get-Content read from a URL? Never tried to read a 1.5GB file with that function, I would think it would be to slow at minimum. I would've start with |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi! I'm trying to load a big (~1.5GB) html file in old "excel 2003 xmlns" format to xlsx w/o Excel.
For "smaller" files that works ok. For large files I'm getting [Get-Content], OutOfMemoryException.
Is there anyway to solve that?
function Get-HtmlTable {
param(
[Parameter(Mandatory=$true)]
$Url,
$TableIndex=0,
$Header,
[int]$FirstDataRow=0,
[Switch]$UseDefaultCredentials
)
}
function Import-Html {
[CmdletBinding()]
param(
$Url,
[int]$Index = 0,
$Header,
[int]$FirstDataRow = 0,
[Switch]$UseDefaultCredentials
)
}
Import-Html $args[0] 0
Beta Was this translation helpful? Give feedback.
All reactions