I've been looking for some way to dynamically reference another workbook of excel using concatenate and indirect. And just get more and more frustrated.
This is how my problem looks like.
Concatenate inside Indirect formula
How do I make my formula in cell 'B3' work?
From the picture it appears that you are missing an apostrophe at the beginning of your reference. It should look like this:
'D:\path\[Filename.xlsx]sheet 1'!$D$6
#REF! might as well be caused by just wrong filename. Please check it.
And finally: the referenced file must be open for this formula to work. Do not use INDIRECT() to extract data from other files stored somewhere in system - this is not possible and should be avoided because file references generated from text strings cannot be tracked from Data / Edit Links menu. You should consider using regular formula links to other files instead - you can easily refresh your references from said menu.
Related
I am attempting to understand the 2020q1 data set found here: https://www.sec.gov/dera/data/financial-statement-data-sets.html,
and am using the reference documentation inside the 2020q1 folder as a “readme” file. The reference documentation specifies that
within the Presentation (pre) data set, the “report” field is a numeric (integer) whose “value refers to the “R file” as posted
on the EDGAR Web site.” I have found no such file after extensive search, and am left with no method of interpreting the “report”
field and all associated data. Please link to the appropriate R File or guide me in the right direction for assistance if possible. Thanks!
So a point of clarification upfront, cause this confused me as well, the "R file" in question is not a script file of the R language. Instead, it simply seems to be a report file that holds the formatted data.
So after digging deeper into the readme, I found the following detail in the description for the SUB.txt data.
Note: To access the complete submission files for a given filing, please see the SEC EDGAR website. The SEC website folder HTTP(s)://www.sec.gov/Archives/edgar/data/{cik}/{accession}/ will always contain all the data sets for a given submission. To assemble the folder address to any filing referenced in the SUB data set, simply substitute {cik} with the cik field and replace {accession} with the adsh field (after removing the dash character). The following sample SQL Query provides an example of how to generate a list of addresses for filings contained in the SUB data set:
· select name,form,period, 'http(s)://www.sec.gov/Archives/edgar/data/' + ltrim(str(cik,10))+'/' + replace(adsh,'-','')+'/'+instance as url from SUBM subm order by period desc, name
Therefore, it looks like we have to correlate each "adsh" submission ID with the "cik" company ID in order to get the link we are looking for.
Doing this for the first entry of pre.txt, we get an adsh value of "0001032208-20-000006". I simply searched through sub.txt with notepad and found its associated cik of "1032208" which belongs to "SEMPRA ENERGY". Therefore, we generate the following link: http://www.sec.gov/Archives/edgar/data/1032208/000103220820000006
From there, we find a directory of files associated with the given submission. Inside is a collection of files with the prefix of "R". Simply clicking on them will open them in your browser, using the "report" and "line" fields, we can then correlate which file we want. Notice that we can add "/R{number}.htm" at the end of the link we generated to find this folder to get a given report number.
If you know what you are looking for, doing this by hand with "ctr+f" find functionality should be fine. Otherwise, you may want to open these docs in excel to generate the links for you.
In my Foreach loop I only want to process files that have crosswlk in the filename. Unfortunately this is in the middle of the filename and I can't do crosswlk.pgp.
I have Googled this a bunch and most solutions say to do some C# in a script task to accomplish this. It seems like I should be able to do this in the Foreach container, maybe in the Collection area with Expressions. But I'm having troubles finding any information about the various Properties in the Property Expressions Editor. What's the difference between the Name, FileSpec and FileNameRetrieval properties?
I thought I could do something like:
Property: Name
Expression: `FINDSTRING(#[User::SourceFilename],"crosswlk",1) > 0`
To me that's saying only process files where the filename has crosswlk in it. But that's not working. I did notice the word crosswlk always starts at the 13th character in the filename so I tried using substring, but no luck there either.
Is there a way to do filter the filenames this (or another) way, or is using C# the only way? Thanks very much!
According to the documentation, you should be able to use wildcards in the Files field.
Use wildcard characters (*) to specify the files to include in the
collection. For example, to include files with names that contain
“abc”, use the following filter: *abc*.
So, you should be able to use something like this: *crosswlk*
See here for more detail: https://learn.microsoft.com/en-us/sql/integration-services/control-flow/foreach-loop-container#enumerator-dynamic-options
If it is too hard to craft a usable wildcard then one option that might be of use is to use a FOREACH ADO Enumerator. First you "fill" a recordset Object Variable with all the filenames you want then you can loop through this.
You can fill the recordset object through a Script Task and use C# system.IO for interrogating the filesystem.
This is a handy way if you need to tightly control what files are processed (eg to prevent test files being loaded)
What I'm trying to do should be quite simple but I am getting error messages saying that conditional formatting doesn't accept this method.
So I am looking for a fix or an alternative.
What I have is a list of search texts for phone faults
screen broken
wifi not working
bluetooth dead
the dooflip isnt working
etc.
I have also set up a number of common keywords
screen
wifi
bluetooth
By comparing these two we can see that the last one doesn't adhere to any of my common keywords, and so I want it highlighted/ copied somewhere etc. so I can view and take any action.
So far I have attempted to set up conditional formatting to search the terms for the keywords and if it doesn't contain any then highlight it.
=ISERROR(SEARCH({"wifi", "bluetooth", "screen"},$H$117))
When using this I get informed that conditional formatting does not support arrays (it works fine if I only specify one search term).
If i create a table with all the keywords in and use the following it works, however it will only find direct mataches, and wont search for the word in the cell string
=ISERROR(MATCH(<cell_to_change_colour_of>, <keyword_table>,0))
So is there another way of doing this?
my answer requires MOREFUNC addon*
=REGEX.COMP(A1,"wifi|bluetooth|screen",FALSE)
this will return true if any of the words appear anywhere in cell A1. FALSE in the formula stipulates that it is not case sensitive.
Unfortunately due to Excell's stupid rules it will not work directly in Conditional Formatting box. You will need to place it in a helper column, and point the conditional formatting rule to that column.
MOREFUNC ADDON
Morefunc Addon is a free library of 66 new worksheet functions.
HERE is some information (by original author)
here is the last working download link I found
here is a good installation walk-through video
I am under Windows and I am using Sublime2 Text Editor (i can download and use any other software) or PHP script.
Now, I am searching for solution to advance merge two files (or lines at one file, no matter). KEY - is the same part of two files
sourcefileone.txt:
KEY|CCC
sortcefiletwo.txt
KEY|BBB
Need to merge and receive this:
KEY|BBB|CCC
Any Solution? thanks
seems there is no ready solution. To solve the problem I already wrote php code that load to files in array, than explode them and merge in the new format at new array.
I want to read a single file (the file is a html document) from my computer and store it in a Corpus (I'm using the package tm).
Do you have any solution to do that?
Here is what I tried :
data<-read.csv(fileName)
c2<-Corpus(VectorSource(data))
it mostly works, but I sometime get the error : more columns than column names
I guess I'm not supposed to use read.csv for a webpage, as I didn't find a better solution.
Thanks for your help =)
A webpage definitely does not conform to the specifications that a CSV should. Instead you probably want to use the readHTMLTable function from the XML package.
This is grabbing from an actual webpage but it should be the same idea
file <- "http://xkcd.com/"
dat <- readLines(file)
c2 <- Corpus(VectorSource(dat))